* Change graceful to use a singleton obtained through GetManager instead of a global. * Graceful: Make TestPullRequests shutdownable * Graceful: Make the cron tasks graceful * Graceful: AddTestPullRequest run in graceful ctx * Graceful: SyncMirrors shutdown * Graceful: SetDefaultContext for Xorm to be HammerContext * Avoid starting graceful for migrate commands and checkout * Graceful: DeliverHooks now can be shutdown * Fix multiple syncing errors in modules/sync/UniqueQueue & Make UniqueQueue closable * Begin the process of making the repo indexer shutdown gracefullytags/v1.21.12.1
| @@ -5,6 +5,8 @@ | |||||
| package cmd | package cmd | ||||
| import ( | import ( | ||||
| "context" | |||||
| "code.gitea.io/gitea/models" | "code.gitea.io/gitea/models" | ||||
| "code.gitea.io/gitea/models/migrations" | "code.gitea.io/gitea/models/migrations" | ||||
| "code.gitea.io/gitea/modules/log" | "code.gitea.io/gitea/modules/log" | ||||
| @@ -32,7 +34,7 @@ func runMigrate(ctx *cli.Context) error { | |||||
| log.Trace("Log path: %s", setting.LogRootPath) | log.Trace("Log path: %s", setting.LogRootPath) | ||||
| setting.InitDBConfig() | setting.InitDBConfig() | ||||
| if err := models.NewEngine(migrations.Migrate); err != nil { | |||||
| if err := models.NewEngine(context.Background(), migrations.Migrate); err != nil { | |||||
| log.Fatal("Failed to initialize ORM engine: %v", err) | log.Fatal("Failed to initialize ORM engine: %v", err) | ||||
| return err | return err | ||||
| } | } | ||||
| @@ -5,6 +5,7 @@ | |||||
| package cmd | package cmd | ||||
| import ( | import ( | ||||
| "context" | |||||
| "fmt" | "fmt" | ||||
| "net/http" | "net/http" | ||||
| _ "net/http/pprof" // Used for debugging if enabled and a web server is running | _ "net/http/pprof" // Used for debugging if enabled and a web server is running | ||||
| @@ -96,6 +97,10 @@ func runLetsEncryptFallbackHandler(w http.ResponseWriter, r *http.Request) { | |||||
| } | } | ||||
| func runWeb(ctx *cli.Context) error { | func runWeb(ctx *cli.Context) error { | ||||
| managerCtx, cancel := context.WithCancel(context.Background()) | |||||
| graceful.InitManager(managerCtx) | |||||
| defer cancel() | |||||
| if os.Getppid() > 1 && len(os.Getenv("LISTEN_FDS")) > 0 { | if os.Getppid() > 1 && len(os.Getenv("LISTEN_FDS")) > 0 { | ||||
| log.Info("Restarting Gitea on PID: %d from parent PID: %d", os.Getpid(), os.Getppid()) | log.Info("Restarting Gitea on PID: %d from parent PID: %d", os.Getpid(), os.Getppid()) | ||||
| } else { | } else { | ||||
| @@ -108,7 +113,7 @@ func runWeb(ctx *cli.Context) error { | |||||
| } | } | ||||
| // Perform global initialization | // Perform global initialization | ||||
| routers.GlobalInit() | |||||
| routers.GlobalInit(graceful.GetManager().HammerContext()) | |||||
| // Set up Macaron | // Set up Macaron | ||||
| m := routes.NewMacaron() | m := routes.NewMacaron() | ||||
| @@ -199,8 +204,7 @@ func runWeb(ctx *cli.Context) error { | |||||
| log.Critical("Failed to start server: %v", err) | log.Critical("Failed to start server: %v", err) | ||||
| } | } | ||||
| log.Info("HTTP Listener: %s Closed", listenAddr) | log.Info("HTTP Listener: %s Closed", listenAddr) | ||||
| graceful.Manager.WaitForServers() | |||||
| graceful.Manager.WaitForTerminate() | |||||
| <-graceful.GetManager().Done() | |||||
| log.Info("PID: %d Gitea Web Finished", os.Getpid()) | log.Info("PID: %d Gitea Web Finished", os.Getpid()) | ||||
| log.Close() | log.Close() | ||||
| return nil | return nil | ||||
| @@ -28,13 +28,13 @@ func runHTTPSWithTLSConfig(network, listenAddr string, tlsConfig *tls.Config, m | |||||
| // NoHTTPRedirector tells our cleanup routine that we will not be using a fallback http redirector | // NoHTTPRedirector tells our cleanup routine that we will not be using a fallback http redirector | ||||
| func NoHTTPRedirector() { | func NoHTTPRedirector() { | ||||
| graceful.Manager.InformCleanup() | |||||
| graceful.GetManager().InformCleanup() | |||||
| } | } | ||||
| // NoMainListener tells our cleanup routine that we will not be using a possibly provided listener | // NoMainListener tells our cleanup routine that we will not be using a possibly provided listener | ||||
| // for our main HTTP/HTTPS service | // for our main HTTP/HTTPS service | ||||
| func NoMainListener() { | func NoMainListener() { | ||||
| graceful.Manager.InformCleanup() | |||||
| graceful.GetManager().InformCleanup() | |||||
| } | } | ||||
| func runFCGI(network, listenAddr string, m http.Handler) error { | func runFCGI(network, listenAddr string, m http.Handler) error { | ||||
| @@ -5,6 +5,7 @@ Checkout a PR and load the tests data into sqlite database | |||||
| */ | */ | ||||
| import ( | import ( | ||||
| "context" | |||||
| "flag" | "flag" | ||||
| "fmt" | "fmt" | ||||
| "io/ioutil" | "io/ioutil" | ||||
| @@ -92,7 +93,7 @@ func runPR() { | |||||
| //x, err = xorm.NewEngine("sqlite3", "file::memory:?cache=shared") | //x, err = xorm.NewEngine("sqlite3", "file::memory:?cache=shared") | ||||
| var helper testfixtures.Helper = &testfixtures.SQLite{} | var helper testfixtures.Helper = &testfixtures.SQLite{} | ||||
| models.NewEngine(func(_ *xorm.Engine) error { | |||||
| models.NewEngine(context.Background(), func(_ *xorm.Engine) error { | |||||
| return nil | return nil | ||||
| }) | }) | ||||
| models.HasEngine = true | models.HasEngine = true | ||||
| @@ -5,6 +5,7 @@ | |||||
| package integrations | package integrations | ||||
| import ( | import ( | ||||
| "context" | |||||
| "net/http" | "net/http" | ||||
| "os" | "os" | ||||
| "strings" | "strings" | ||||
| @@ -147,7 +148,7 @@ func TestLDAPUserSync(t *testing.T) { | |||||
| } | } | ||||
| defer prepareTestEnv(t)() | defer prepareTestEnv(t)() | ||||
| addAuthSourceLDAP(t, "") | addAuthSourceLDAP(t, "") | ||||
| models.SyncExternalUsers() | |||||
| models.SyncExternalUsers(context.Background()) | |||||
| session := loginUser(t, "user1") | session := loginUser(t, "user1") | ||||
| // Check if users exists | // Check if users exists | ||||
| @@ -206,7 +207,8 @@ func TestLDAPUserSSHKeySync(t *testing.T) { | |||||
| } | } | ||||
| defer prepareTestEnv(t)() | defer prepareTestEnv(t)() | ||||
| addAuthSourceLDAP(t, "sshPublicKey") | addAuthSourceLDAP(t, "sshPublicKey") | ||||
| models.SyncExternalUsers() | |||||
| models.SyncExternalUsers(context.Background()) | |||||
| // Check if users has SSH keys synced | // Check if users has SSH keys synced | ||||
| for _, u := range gitLDAPUsers { | for _, u := range gitLDAPUsers { | ||||
| @@ -6,6 +6,7 @@ package integrations | |||||
| import ( | import ( | ||||
| "bytes" | "bytes" | ||||
| "context" | |||||
| "database/sql" | "database/sql" | ||||
| "encoding/json" | "encoding/json" | ||||
| "fmt" | "fmt" | ||||
| @@ -24,6 +25,7 @@ import ( | |||||
| "code.gitea.io/gitea/models" | "code.gitea.io/gitea/models" | ||||
| "code.gitea.io/gitea/modules/base" | "code.gitea.io/gitea/modules/base" | ||||
| "code.gitea.io/gitea/modules/graceful" | |||||
| "code.gitea.io/gitea/modules/setting" | "code.gitea.io/gitea/modules/setting" | ||||
| "code.gitea.io/gitea/routers" | "code.gitea.io/gitea/routers" | ||||
| "code.gitea.io/gitea/routers/routes" | "code.gitea.io/gitea/routers/routes" | ||||
| @@ -55,6 +57,10 @@ func NewNilResponseRecorder() *NilResponseRecorder { | |||||
| } | } | ||||
| func TestMain(m *testing.M) { | func TestMain(m *testing.M) { | ||||
| managerCtx, cancel := context.WithCancel(context.Background()) | |||||
| graceful.InitManager(managerCtx) | |||||
| defer cancel() | |||||
| initIntegrationTest() | initIntegrationTest() | ||||
| mac = routes.NewMacaron() | mac = routes.NewMacaron() | ||||
| routes.RegisterRoutes(mac) | routes.RegisterRoutes(mac) | ||||
| @@ -171,7 +177,7 @@ func initIntegrationTest() { | |||||
| } | } | ||||
| defer db.Close() | defer db.Close() | ||||
| } | } | ||||
| routers.GlobalInit() | |||||
| routers.GlobalInit(graceful.GetManager().HammerContext()) | |||||
| } | } | ||||
| func prepareTestEnv(t testing.TB, skip ...int) func() { | func prepareTestEnv(t testing.TB, skip ...int) func() { | ||||
| @@ -6,6 +6,7 @@ package migrations | |||||
| import ( | import ( | ||||
| "compress/gzip" | "compress/gzip" | ||||
| "context" | |||||
| "database/sql" | "database/sql" | ||||
| "fmt" | "fmt" | ||||
| "io/ioutil" | "io/ioutil" | ||||
| @@ -220,7 +221,7 @@ func doMigrationTest(t *testing.T, version string) { | |||||
| err := models.SetEngine() | err := models.SetEngine() | ||||
| assert.NoError(t, err) | assert.NoError(t, err) | ||||
| err = models.NewEngine(wrappedMigrate) | |||||
| err = models.NewEngine(context.Background(), wrappedMigrate) | |||||
| assert.NoError(t, err) | assert.NoError(t, err) | ||||
| currentEngine.Close() | currentEngine.Close() | ||||
| } | } | ||||
| @@ -5,6 +5,7 @@ | |||||
| package models | package models | ||||
| import ( | import ( | ||||
| "context" | |||||
| "fmt" | "fmt" | ||||
| "time" | "time" | ||||
| @@ -525,7 +526,8 @@ func (deletedBranch *DeletedBranch) LoadUser() { | |||||
| } | } | ||||
| // RemoveOldDeletedBranches removes old deleted branches | // RemoveOldDeletedBranches removes old deleted branches | ||||
| func RemoveOldDeletedBranches() { | |||||
| func RemoveOldDeletedBranches(ctx context.Context) { | |||||
| // Nothing to do for shutdown or terminate | |||||
| log.Trace("Doing: DeletedBranchesCleanup") | log.Trace("Doing: DeletedBranchesCleanup") | ||||
| deleteBefore := time.Now().Add(-setting.Cron.DeletedBranchesCleanup.OlderThan) | deleteBefore := time.Now().Add(-setting.Cron.DeletedBranchesCleanup.OlderThan) | ||||
| @@ -6,6 +6,7 @@ | |||||
| package models | package models | ||||
| import ( | import ( | ||||
| "context" | |||||
| "database/sql" | "database/sql" | ||||
| "errors" | "errors" | ||||
| "fmt" | "fmt" | ||||
| @@ -164,11 +165,13 @@ func SetEngine() (err error) { | |||||
| } | } | ||||
| // NewEngine initializes a new xorm.Engine | // NewEngine initializes a new xorm.Engine | ||||
| func NewEngine(migrateFunc func(*xorm.Engine) error) (err error) { | |||||
| func NewEngine(ctx context.Context, migrateFunc func(*xorm.Engine) error) (err error) { | |||||
| if err = SetEngine(); err != nil { | if err = SetEngine(); err != nil { | ||||
| return err | return err | ||||
| } | } | ||||
| x.SetDefaultContext(ctx) | |||||
| if err = x.Ping(); err != nil { | if err = x.Ping(); err != nil { | ||||
| return err | return err | ||||
| } | } | ||||
| @@ -68,11 +68,12 @@ func GetUnmergedPullRequestsByBaseInfo(repoID int64, branch string) ([]*PullRequ | |||||
| Find(&prs) | Find(&prs) | ||||
| } | } | ||||
| // GetPullRequestsByCheckStatus returns all pull requests according the special checking status. | |||||
| func GetPullRequestsByCheckStatus(status PullRequestStatus) ([]*PullRequest, error) { | |||||
| prs := make([]*PullRequest, 0, 10) | |||||
| return prs, x. | |||||
| // GetPullRequestIDsByCheckStatus returns all pull requests according the special checking status. | |||||
| func GetPullRequestIDsByCheckStatus(status PullRequestStatus) ([]int64, error) { | |||||
| prs := make([]int64, 0, 10) | |||||
| return prs, x.Table("pull_request"). | |||||
| Where("status=?", status). | Where("status=?", status). | ||||
| Cols("pull_request.id"). | |||||
| Find(&prs) | Find(&prs) | ||||
| } | } | ||||
| @@ -7,6 +7,7 @@ package models | |||||
| import ( | import ( | ||||
| "bytes" | "bytes" | ||||
| "context" | |||||
| "crypto/md5" | "crypto/md5" | ||||
| "errors" | "errors" | ||||
| "fmt" | "fmt" | ||||
| @@ -2098,19 +2099,27 @@ func DeleteRepositoryArchives() error { | |||||
| } | } | ||||
| // DeleteOldRepositoryArchives deletes old repository archives. | // DeleteOldRepositoryArchives deletes old repository archives. | ||||
| func DeleteOldRepositoryArchives() { | |||||
| func DeleteOldRepositoryArchives(ctx context.Context) { | |||||
| log.Trace("Doing: ArchiveCleanup") | log.Trace("Doing: ArchiveCleanup") | ||||
| if err := x.Where("id > 0").Iterate(new(Repository), deleteOldRepositoryArchives); err != nil { | |||||
| if err := x.Where("id > 0").Iterate(new(Repository), func(idx int, bean interface{}) error { | |||||
| return deleteOldRepositoryArchives(ctx, idx, bean) | |||||
| }); err != nil { | |||||
| log.Error("ArchiveClean: %v", err) | log.Error("ArchiveClean: %v", err) | ||||
| } | } | ||||
| } | } | ||||
| func deleteOldRepositoryArchives(idx int, bean interface{}) error { | |||||
| func deleteOldRepositoryArchives(ctx context.Context, idx int, bean interface{}) error { | |||||
| repo := bean.(*Repository) | repo := bean.(*Repository) | ||||
| basePath := filepath.Join(repo.RepoPath(), "archives") | basePath := filepath.Join(repo.RepoPath(), "archives") | ||||
| for _, ty := range []string{"zip", "targz"} { | for _, ty := range []string{"zip", "targz"} { | ||||
| select { | |||||
| case <-ctx.Done(): | |||||
| return fmt.Errorf("Aborted due to shutdown:\nin delete of old repository archives %v\nat delete file %s", repo, ty) | |||||
| default: | |||||
| } | |||||
| path := filepath.Join(basePath, ty) | path := filepath.Join(basePath, ty) | ||||
| file, err := os.Open(path) | file, err := os.Open(path) | ||||
| if err != nil { | if err != nil { | ||||
| @@ -2133,6 +2142,11 @@ func deleteOldRepositoryArchives(idx int, bean interface{}) error { | |||||
| minimumOldestTime := time.Now().Add(-setting.Cron.ArchiveCleanup.OlderThan) | minimumOldestTime := time.Now().Add(-setting.Cron.ArchiveCleanup.OlderThan) | ||||
| for _, info := range files { | for _, info := range files { | ||||
| if info.ModTime().Before(minimumOldestTime) && !info.IsDir() { | if info.ModTime().Before(minimumOldestTime) && !info.IsDir() { | ||||
| select { | |||||
| case <-ctx.Done(): | |||||
| return fmt.Errorf("Aborted due to shutdown:\nin delete of old repository archives %v\nat delete file %s - %s", repo, ty, info.Name()) | |||||
| default: | |||||
| } | |||||
| toDelete := filepath.Join(path, info.Name()) | toDelete := filepath.Join(path, info.Name()) | ||||
| // This is a best-effort purge, so we do not check error codes to confirm removal. | // This is a best-effort purge, so we do not check error codes to confirm removal. | ||||
| if err = os.Remove(toDelete); err != nil { | if err = os.Remove(toDelete); err != nil { | ||||
| @@ -2226,13 +2240,17 @@ func SyncRepositoryHooks() error { | |||||
| } | } | ||||
| // GitFsck calls 'git fsck' to check repository health. | // GitFsck calls 'git fsck' to check repository health. | ||||
| func GitFsck() { | |||||
| func GitFsck(ctx context.Context) { | |||||
| log.Trace("Doing: GitFsck") | log.Trace("Doing: GitFsck") | ||||
| if err := x. | if err := x. | ||||
| Where("id>0 AND is_fsck_enabled=?", true).BufferSize(setting.Database.IterateBufferSize). | Where("id>0 AND is_fsck_enabled=?", true).BufferSize(setting.Database.IterateBufferSize). | ||||
| Iterate(new(Repository), | Iterate(new(Repository), | ||||
| func(idx int, bean interface{}) error { | func(idx int, bean interface{}) error { | ||||
| select { | |||||
| case <-ctx.Done(): | |||||
| return fmt.Errorf("Aborted due to shutdown") | |||||
| default: | |||||
| } | |||||
| repo := bean.(*Repository) | repo := bean.(*Repository) | ||||
| repoPath := repo.RepoPath() | repoPath := repo.RepoPath() | ||||
| log.Trace("Running health check on repository %s", repoPath) | log.Trace("Running health check on repository %s", repoPath) | ||||
| @@ -2278,13 +2296,19 @@ type repoChecker struct { | |||||
| desc string | desc string | ||||
| } | } | ||||
| func repoStatsCheck(checker *repoChecker) { | |||||
| func repoStatsCheck(ctx context.Context, checker *repoChecker) { | |||||
| results, err := x.Query(checker.querySQL) | results, err := x.Query(checker.querySQL) | ||||
| if err != nil { | if err != nil { | ||||
| log.Error("Select %s: %v", checker.desc, err) | log.Error("Select %s: %v", checker.desc, err) | ||||
| return | return | ||||
| } | } | ||||
| for _, result := range results { | for _, result := range results { | ||||
| select { | |||||
| case <-ctx.Done(): | |||||
| log.Warn("CheckRepoStats: Aborting due to shutdown") | |||||
| return | |||||
| default: | |||||
| } | |||||
| id := com.StrTo(result["id"]).MustInt64() | id := com.StrTo(result["id"]).MustInt64() | ||||
| log.Trace("Updating %s: %d", checker.desc, id) | log.Trace("Updating %s: %d", checker.desc, id) | ||||
| _, err = x.Exec(checker.correctSQL, id, id) | _, err = x.Exec(checker.correctSQL, id, id) | ||||
| @@ -2295,7 +2319,7 @@ func repoStatsCheck(checker *repoChecker) { | |||||
| } | } | ||||
| // CheckRepoStats checks the repository stats | // CheckRepoStats checks the repository stats | ||||
| func CheckRepoStats() { | |||||
| func CheckRepoStats(ctx context.Context) { | |||||
| log.Trace("Doing: CheckRepoStats") | log.Trace("Doing: CheckRepoStats") | ||||
| checkers := []*repoChecker{ | checkers := []*repoChecker{ | ||||
| @@ -2331,7 +2355,13 @@ func CheckRepoStats() { | |||||
| }, | }, | ||||
| } | } | ||||
| for i := range checkers { | for i := range checkers { | ||||
| repoStatsCheck(checkers[i]) | |||||
| select { | |||||
| case <-ctx.Done(): | |||||
| log.Warn("CheckRepoStats: Aborting due to shutdown") | |||||
| return | |||||
| default: | |||||
| repoStatsCheck(ctx, checkers[i]) | |||||
| } | |||||
| } | } | ||||
| // ***** START: Repository.NumClosedIssues ***** | // ***** START: Repository.NumClosedIssues ***** | ||||
| @@ -2341,6 +2371,12 @@ func CheckRepoStats() { | |||||
| log.Error("Select %s: %v", desc, err) | log.Error("Select %s: %v", desc, err) | ||||
| } else { | } else { | ||||
| for _, result := range results { | for _, result := range results { | ||||
| select { | |||||
| case <-ctx.Done(): | |||||
| log.Warn("CheckRepoStats: Aborting due to shutdown") | |||||
| return | |||||
| default: | |||||
| } | |||||
| id := com.StrTo(result["id"]).MustInt64() | id := com.StrTo(result["id"]).MustInt64() | ||||
| log.Trace("Updating %s: %d", desc, id) | log.Trace("Updating %s: %d", desc, id) | ||||
| _, err = x.Exec("UPDATE `repository` SET num_closed_issues=(SELECT COUNT(*) FROM `issue` WHERE repo_id=? AND is_closed=? AND is_pull=?) WHERE id=?", id, true, false, id) | _, err = x.Exec("UPDATE `repository` SET num_closed_issues=(SELECT COUNT(*) FROM `issue` WHERE repo_id=? AND is_closed=? AND is_pull=?) WHERE id=?", id, true, false, id) | ||||
| @@ -2358,6 +2394,12 @@ func CheckRepoStats() { | |||||
| log.Error("Select %s: %v", desc, err) | log.Error("Select %s: %v", desc, err) | ||||
| } else { | } else { | ||||
| for _, result := range results { | for _, result := range results { | ||||
| select { | |||||
| case <-ctx.Done(): | |||||
| log.Warn("CheckRepoStats: Aborting due to shutdown") | |||||
| return | |||||
| default: | |||||
| } | |||||
| id := com.StrTo(result["id"]).MustInt64() | id := com.StrTo(result["id"]).MustInt64() | ||||
| log.Trace("Updating %s: %d", desc, id) | log.Trace("Updating %s: %d", desc, id) | ||||
| _, err = x.Exec("UPDATE `repository` SET num_closed_pulls=(SELECT COUNT(*) FROM `issue` WHERE repo_id=? AND is_closed=? AND is_pull=?) WHERE id=?", id, true, true, id) | _, err = x.Exec("UPDATE `repository` SET num_closed_pulls=(SELECT COUNT(*) FROM `issue` WHERE repo_id=? AND is_closed=? AND is_pull=?) WHERE id=?", id, true, true, id) | ||||
| @@ -2375,6 +2417,12 @@ func CheckRepoStats() { | |||||
| log.Error("Select repository count 'num_forks': %v", err) | log.Error("Select repository count 'num_forks': %v", err) | ||||
| } else { | } else { | ||||
| for _, result := range results { | for _, result := range results { | ||||
| select { | |||||
| case <-ctx.Done(): | |||||
| log.Warn("CheckRepoStats: Aborting due to shutdown") | |||||
| return | |||||
| default: | |||||
| } | |||||
| id := com.StrTo(result["id"]).MustInt64() | id := com.StrTo(result["id"]).MustInt64() | ||||
| log.Trace("Updating repository count 'num_forks': %d", id) | log.Trace("Updating repository count 'num_forks': %d", id) | ||||
| @@ -4,6 +4,12 @@ | |||||
| package models | package models | ||||
| import ( | |||||
| "fmt" | |||||
| "xorm.io/builder" | |||||
| ) | |||||
| // RepoIndexerStatus status of a repo's entry in the repo indexer | // RepoIndexerStatus status of a repo's entry in the repo indexer | ||||
| // For now, implicitly refers to default branch | // For now, implicitly refers to default branch | ||||
| type RepoIndexerStatus struct { | type RepoIndexerStatus struct { | ||||
| @@ -12,6 +18,31 @@ type RepoIndexerStatus struct { | |||||
| CommitSha string `xorm:"VARCHAR(40)"` | CommitSha string `xorm:"VARCHAR(40)"` | ||||
| } | } | ||||
| // GetUnindexedRepos returns repos which do not have an indexer status | |||||
| func GetUnindexedRepos(maxRepoID int64, page, pageSize int) ([]int64, error) { | |||||
| ids := make([]int64, 0, 50) | |||||
| cond := builder.Cond(builder.IsNull{ | |||||
| "repo_indexer_status.id", | |||||
| }) | |||||
| sess := x.Table("repository").Join("LEFT OUTER", "repo_indexer_status", "repository.id = repo_indexer_status.repo_id") | |||||
| if maxRepoID > 0 { | |||||
| cond = builder.And(cond, builder.Lte{ | |||||
| "repository.id": maxRepoID, | |||||
| }) | |||||
| } | |||||
| if page >= 0 && pageSize > 0 { | |||||
| start := 0 | |||||
| if page > 0 { | |||||
| start = (page - 1) * pageSize | |||||
| } | |||||
| sess.Limit(pageSize, start) | |||||
| } | |||||
| sess.Where(cond).Cols("repository.id").Desc("repository.id") | |||||
| err := sess.Find(&ids) | |||||
| return ids, err | |||||
| } | |||||
| // GetIndexerStatus loads repo codes indxer status | // GetIndexerStatus loads repo codes indxer status | ||||
| func (repo *Repository) GetIndexerStatus() error { | func (repo *Repository) GetIndexerStatus() error { | ||||
| if repo.IndexerStatus != nil { | if repo.IndexerStatus != nil { | ||||
| @@ -31,15 +62,21 @@ func (repo *Repository) GetIndexerStatus() error { | |||||
| // UpdateIndexerStatus updates indexer status | // UpdateIndexerStatus updates indexer status | ||||
| func (repo *Repository) UpdateIndexerStatus(sha string) error { | func (repo *Repository) UpdateIndexerStatus(sha string) error { | ||||
| if err := repo.GetIndexerStatus(); err != nil { | if err := repo.GetIndexerStatus(); err != nil { | ||||
| return err | |||||
| return fmt.Errorf("UpdateIndexerStatus: Unable to getIndexerStatus for repo: %s/%s Error: %v", repo.MustOwnerName(), repo.Name, err) | |||||
| } | } | ||||
| if len(repo.IndexerStatus.CommitSha) == 0 { | if len(repo.IndexerStatus.CommitSha) == 0 { | ||||
| repo.IndexerStatus.CommitSha = sha | repo.IndexerStatus.CommitSha = sha | ||||
| _, err := x.Insert(repo.IndexerStatus) | _, err := x.Insert(repo.IndexerStatus) | ||||
| return err | |||||
| if err != nil { | |||||
| return fmt.Errorf("UpdateIndexerStatus: Unable to insert repoIndexerStatus for repo: %s/%s Sha: %s Error: %v", repo.MustOwnerName(), repo.Name, sha, err) | |||||
| } | |||||
| return nil | |||||
| } | } | ||||
| repo.IndexerStatus.CommitSha = sha | repo.IndexerStatus.CommitSha = sha | ||||
| _, err := x.ID(repo.IndexerStatus.ID).Cols("commit_sha"). | _, err := x.ID(repo.IndexerStatus.ID).Cols("commit_sha"). | ||||
| Update(repo.IndexerStatus) | Update(repo.IndexerStatus) | ||||
| return err | |||||
| if err != nil { | |||||
| return fmt.Errorf("UpdateIndexerStatus: Unable to update repoIndexerStatus for repo: %s/%s Sha: %s Error: %v", repo.MustOwnerName(), repo.Name, sha, err) | |||||
| } | |||||
| return nil | |||||
| } | } | ||||
| @@ -7,6 +7,7 @@ package models | |||||
| import ( | import ( | ||||
| "container/list" | "container/list" | ||||
| "context" | |||||
| "crypto/md5" | "crypto/md5" | ||||
| "crypto/sha256" | "crypto/sha256" | ||||
| "crypto/subtle" | "crypto/subtle" | ||||
| @@ -1695,7 +1696,7 @@ func synchronizeLdapSSHPublicKeys(usr *User, s *LoginSource, sshPublicKeys []str | |||||
| } | } | ||||
| // SyncExternalUsers is used to synchronize users with external authorization source | // SyncExternalUsers is used to synchronize users with external authorization source | ||||
| func SyncExternalUsers() { | |||||
| func SyncExternalUsers(ctx context.Context) { | |||||
| log.Trace("Doing: SyncExternalUsers") | log.Trace("Doing: SyncExternalUsers") | ||||
| ls, err := LoginSources() | ls, err := LoginSources() | ||||
| @@ -1710,6 +1711,12 @@ func SyncExternalUsers() { | |||||
| if !s.IsActived || !s.IsSyncEnabled { | if !s.IsActived || !s.IsSyncEnabled { | ||||
| continue | continue | ||||
| } | } | ||||
| select { | |||||
| case <-ctx.Done(): | |||||
| log.Warn("SyncExternalUsers: Aborted due to shutdown before update of %s", s.Name) | |||||
| return | |||||
| default: | |||||
| } | |||||
| if s.IsLDAP() { | if s.IsLDAP() { | ||||
| log.Trace("Doing: SyncExternalUsers[%s]", s.Name) | log.Trace("Doing: SyncExternalUsers[%s]", s.Name) | ||||
| @@ -1727,6 +1734,12 @@ func SyncExternalUsers() { | |||||
| log.Error("SyncExternalUsers: %v", err) | log.Error("SyncExternalUsers: %v", err) | ||||
| return | return | ||||
| } | } | ||||
| select { | |||||
| case <-ctx.Done(): | |||||
| log.Warn("SyncExternalUsers: Aborted due to shutdown before update of %s", s.Name) | |||||
| return | |||||
| default: | |||||
| } | |||||
| sr, err := s.LDAP().SearchEntries() | sr, err := s.LDAP().SearchEntries() | ||||
| if err != nil { | if err != nil { | ||||
| @@ -1735,6 +1748,19 @@ func SyncExternalUsers() { | |||||
| } | } | ||||
| for _, su := range sr { | for _, su := range sr { | ||||
| select { | |||||
| case <-ctx.Done(): | |||||
| log.Warn("SyncExternalUsers: Aborted due to shutdown at update of %s before completed update of users", s.Name) | |||||
| // Rewrite authorized_keys file if LDAP Public SSH Key attribute is set and any key was added or removed | |||||
| if sshKeysNeedUpdate { | |||||
| err = RewriteAllPublicKeys() | |||||
| if err != nil { | |||||
| log.Error("RewriteAllPublicKeys: %v", err) | |||||
| } | |||||
| } | |||||
| return | |||||
| default: | |||||
| } | |||||
| if len(su.Username) == 0 { | if len(su.Username) == 0 { | ||||
| continue | continue | ||||
| } | } | ||||
| @@ -1819,6 +1845,13 @@ func SyncExternalUsers() { | |||||
| } | } | ||||
| } | } | ||||
| select { | |||||
| case <-ctx.Done(): | |||||
| log.Warn("SyncExternalUsers: Aborted due to shutdown at update of %s before delete users", s.Name) | |||||
| return | |||||
| default: | |||||
| } | |||||
| // Deactivate users not present in LDAP | // Deactivate users not present in LDAP | ||||
| if updateExisting { | if updateExisting { | ||||
| for _, usr := range users { | for _, usr := range users { | ||||
| @@ -6,9 +6,11 @@ | |||||
| package cron | package cron | ||||
| import ( | import ( | ||||
| "context" | |||||
| "time" | "time" | ||||
| "code.gitea.io/gitea/models" | "code.gitea.io/gitea/models" | ||||
| "code.gitea.io/gitea/modules/graceful" | |||||
| "code.gitea.io/gitea/modules/log" | "code.gitea.io/gitea/modules/log" | ||||
| "code.gitea.io/gitea/modules/migrations" | "code.gitea.io/gitea/modules/migrations" | ||||
| "code.gitea.io/gitea/modules/setting" | "code.gitea.io/gitea/modules/setting" | ||||
| @@ -37,17 +39,19 @@ var taskStatusTable = sync.NewStatusTable() | |||||
| type Func func() | type Func func() | ||||
| // WithUnique wrap a cron func with an unique running check | // WithUnique wrap a cron func with an unique running check | ||||
| func WithUnique(name string, body Func) Func { | |||||
| func WithUnique(name string, body func(context.Context)) Func { | |||||
| return func() { | return func() { | ||||
| if !taskStatusTable.StartIfNotRunning(name) { | if !taskStatusTable.StartIfNotRunning(name) { | ||||
| return | return | ||||
| } | } | ||||
| defer taskStatusTable.Stop(name) | defer taskStatusTable.Stop(name) | ||||
| body() | |||||
| graceful.GetManager().RunWithShutdownContext(body) | |||||
| } | } | ||||
| } | } | ||||
| // NewContext begins cron tasks | // NewContext begins cron tasks | ||||
| // Each cron task is run within the shutdown context as a running server | |||||
| // AtShutdown the cron server is stopped | |||||
| func NewContext() { | func NewContext() { | ||||
| var ( | var ( | ||||
| entry *cron.Entry | entry *cron.Entry | ||||
| @@ -129,6 +133,7 @@ func NewContext() { | |||||
| go WithUnique(updateMigrationPosterID, migrations.UpdateMigrationPosterID)() | go WithUnique(updateMigrationPosterID, migrations.UpdateMigrationPosterID)() | ||||
| c.Start() | c.Start() | ||||
| graceful.GetManager().RunAtShutdown(context.Background(), c.Stop) | |||||
| } | } | ||||
| // ListTasks returns all running cron tasks. | // ListTasks returns all running cron tasks. | ||||
| @@ -106,7 +106,8 @@ func SetExecutablePath(path string) error { | |||||
| } | } | ||||
| // Init initializes git module | // Init initializes git module | ||||
| func Init() error { | |||||
| func Init(ctx context.Context) error { | |||||
| DefaultContext = ctx | |||||
| // Git requires setting user.name and user.email in order to commit changes. | // Git requires setting user.name and user.email in order to commit changes. | ||||
| for configKey, defaultValue := range map[string]string{"user.name": "Gitea", "user.email": "gitea@fake.local"} { | for configKey, defaultValue := range map[string]string{"user.name": "Gitea", "user.email": "gitea@fake.local"} { | ||||
| if stdout, stderr, err := process.GetManager().Exec("git.Init(get setting)", GitExecutable, "config", "--get", configKey); err != nil || strings.TrimSpace(stdout) == "" { | if stdout, stderr, err := process.GetManager().Exec("git.Init(get setting)", GitExecutable, "config", "--get", configKey); err != nil || strings.TrimSpace(stdout) == "" { | ||||
| @@ -5,6 +5,7 @@ | |||||
| package git | package git | ||||
| import ( | import ( | ||||
| "context" | |||||
| "fmt" | "fmt" | ||||
| "os" | "os" | ||||
| "testing" | "testing" | ||||
| @@ -16,7 +17,7 @@ func fatalTestError(fmtStr string, args ...interface{}) { | |||||
| } | } | ||||
| func TestMain(m *testing.M) { | func TestMain(m *testing.M) { | ||||
| if err := Init(); err != nil { | |||||
| if err := Init(context.Background()); err != nil { | |||||
| fatalTestError("Init failed: %v", err) | fatalTestError("Init failed: %v", err) | ||||
| } | } | ||||
| @@ -62,7 +62,7 @@ func (ctx *ChannelContext) Value(key interface{}) interface{} { | |||||
| // ShutdownContext returns a context.Context that is Done at shutdown | // ShutdownContext returns a context.Context that is Done at shutdown | ||||
| // Callers using this context should ensure that they are registered as a running server | // Callers using this context should ensure that they are registered as a running server | ||||
| // in order that they are waited for. | // in order that they are waited for. | ||||
| func (g *gracefulManager) ShutdownContext() context.Context { | |||||
| func (g *Manager) ShutdownContext() context.Context { | |||||
| return &ChannelContext{ | return &ChannelContext{ | ||||
| done: g.IsShutdown(), | done: g.IsShutdown(), | ||||
| err: ErrShutdown, | err: ErrShutdown, | ||||
| @@ -72,7 +72,7 @@ func (g *gracefulManager) ShutdownContext() context.Context { | |||||
| // HammerContext returns a context.Context that is Done at hammer | // HammerContext returns a context.Context that is Done at hammer | ||||
| // Callers using this context should ensure that they are registered as a running server | // Callers using this context should ensure that they are registered as a running server | ||||
| // in order that they are waited for. | // in order that they are waited for. | ||||
| func (g *gracefulManager) HammerContext() context.Context { | |||||
| func (g *Manager) HammerContext() context.Context { | |||||
| return &ChannelContext{ | return &ChannelContext{ | ||||
| done: g.IsHammer(), | done: g.IsHammer(), | ||||
| err: ErrHammer, | err: ErrHammer, | ||||
| @@ -82,7 +82,7 @@ func (g *gracefulManager) HammerContext() context.Context { | |||||
| // TerminateContext returns a context.Context that is Done at terminate | // TerminateContext returns a context.Context that is Done at terminate | ||||
| // Callers using this context should ensure that they are registered as a terminating server | // Callers using this context should ensure that they are registered as a terminating server | ||||
| // in order that they are waited for. | // in order that they are waited for. | ||||
| func (g *gracefulManager) TerminateContext() context.Context { | |||||
| func (g *Manager) TerminateContext() context.Context { | |||||
| return &ChannelContext{ | return &ChannelContext{ | ||||
| done: g.IsTerminate(), | done: g.IsTerminate(), | ||||
| err: ErrTerminate, | err: ErrTerminate, | ||||
| @@ -6,9 +6,9 @@ package graceful | |||||
| import ( | import ( | ||||
| "context" | "context" | ||||
| "sync" | |||||
| "time" | "time" | ||||
| "code.gitea.io/gitea/modules/git" | |||||
| "code.gitea.io/gitea/modules/log" | "code.gitea.io/gitea/modules/log" | ||||
| "code.gitea.io/gitea/modules/process" | "code.gitea.io/gitea/modules/process" | ||||
| "code.gitea.io/gitea/modules/setting" | "code.gitea.io/gitea/modules/setting" | ||||
| @@ -34,14 +34,24 @@ const ( | |||||
| const numberOfServersToCreate = 3 | const numberOfServersToCreate = 3 | ||||
| // Manager represents the graceful server manager interface | // Manager represents the graceful server manager interface | ||||
| var Manager *gracefulManager | |||||
| func init() { | |||||
| Manager = newGracefulManager(context.Background()) | |||||
| // Set the git default context to the HammerContext | |||||
| git.DefaultContext = Manager.HammerContext() | |||||
| // Set the process default context to the HammerContext | |||||
| process.DefaultContext = Manager.HammerContext() | |||||
| var manager *Manager | |||||
| var initOnce = sync.Once{} | |||||
| // GetManager returns the Manager | |||||
| func GetManager() *Manager { | |||||
| InitManager(context.Background()) | |||||
| return manager | |||||
| } | |||||
| // InitManager creates the graceful manager in the provided context | |||||
| func InitManager(ctx context.Context) { | |||||
| initOnce.Do(func() { | |||||
| manager = newGracefulManager(ctx) | |||||
| // Set the process default context to the HammerContext | |||||
| process.DefaultContext = manager.HammerContext() | |||||
| }) | |||||
| } | } | ||||
| // CallbackWithContext is combined runnable and context to watch to see if the caller has finished | // CallbackWithContext is combined runnable and context to watch to see if the caller has finished | ||||
| @@ -61,7 +71,7 @@ type RunnableWithShutdownFns func(atShutdown, atTerminate func(context.Context, | |||||
| // Please note that use of the atShutdown and atTerminate callbacks will create go-routines that will wait till their respective signals | // Please note that use of the atShutdown and atTerminate callbacks will create go-routines that will wait till their respective signals | ||||
| // - users must therefore be careful to only call these as necessary. | // - users must therefore be careful to only call these as necessary. | ||||
| // If run is not expected to run indefinitely RunWithShutdownChan is likely to be more appropriate. | // If run is not expected to run indefinitely RunWithShutdownChan is likely to be more appropriate. | ||||
| func (g *gracefulManager) RunWithShutdownFns(run RunnableWithShutdownFns) { | |||||
| func (g *Manager) RunWithShutdownFns(run RunnableWithShutdownFns) { | |||||
| g.runningServerWaitGroup.Add(1) | g.runningServerWaitGroup.Add(1) | ||||
| defer g.runningServerWaitGroup.Done() | defer g.runningServerWaitGroup.Done() | ||||
| run(func(ctx context.Context, atShutdown func()) { | run(func(ctx context.Context, atShutdown func()) { | ||||
| @@ -90,7 +100,7 @@ type RunnableWithShutdownChan func(atShutdown <-chan struct{}, atTerminate Callb | |||||
| // (Optionally IsHammer may be waited for instead however, this should be avoided if possible.) | // (Optionally IsHammer may be waited for instead however, this should be avoided if possible.) | ||||
| // The callback function provided to atTerminate must return once termination is complete. | // The callback function provided to atTerminate must return once termination is complete. | ||||
| // Please note that use of the atTerminate function will create a go-routine that will wait till terminate - users must therefore be careful to only call this as necessary. | // Please note that use of the atTerminate function will create a go-routine that will wait till terminate - users must therefore be careful to only call this as necessary. | ||||
| func (g *gracefulManager) RunWithShutdownChan(run RunnableWithShutdownChan) { | |||||
| func (g *Manager) RunWithShutdownChan(run RunnableWithShutdownChan) { | |||||
| g.runningServerWaitGroup.Add(1) | g.runningServerWaitGroup.Add(1) | ||||
| defer g.runningServerWaitGroup.Done() | defer g.runningServerWaitGroup.Done() | ||||
| run(g.IsShutdown(), func(ctx context.Context, atTerminate func()) { | run(g.IsShutdown(), func(ctx context.Context, atTerminate func()) { | ||||
| @@ -101,14 +111,14 @@ func (g *gracefulManager) RunWithShutdownChan(run RunnableWithShutdownChan) { | |||||
| // RunWithShutdownContext takes a function that has a context to watch for shutdown. | // RunWithShutdownContext takes a function that has a context to watch for shutdown. | ||||
| // After the provided context is Done(), the main function must return once shutdown is complete. | // After the provided context is Done(), the main function must return once shutdown is complete. | ||||
| // (Optionally the HammerContext may be obtained and waited for however, this should be avoided if possible.) | // (Optionally the HammerContext may be obtained and waited for however, this should be avoided if possible.) | ||||
| func (g *gracefulManager) RunWithShutdownContext(run func(context.Context)) { | |||||
| func (g *Manager) RunWithShutdownContext(run func(context.Context)) { | |||||
| g.runningServerWaitGroup.Add(1) | g.runningServerWaitGroup.Add(1) | ||||
| defer g.runningServerWaitGroup.Done() | defer g.runningServerWaitGroup.Done() | ||||
| run(g.ShutdownContext()) | run(g.ShutdownContext()) | ||||
| } | } | ||||
| // RunAtTerminate adds to the terminate wait group and creates a go-routine to run the provided function at termination | // RunAtTerminate adds to the terminate wait group and creates a go-routine to run the provided function at termination | ||||
| func (g *gracefulManager) RunAtTerminate(ctx context.Context, terminate func()) { | |||||
| func (g *Manager) RunAtTerminate(ctx context.Context, terminate func()) { | |||||
| g.terminateWaitGroup.Add(1) | g.terminateWaitGroup.Add(1) | ||||
| go func() { | go func() { | ||||
| select { | select { | ||||
| @@ -121,7 +131,7 @@ func (g *gracefulManager) RunAtTerminate(ctx context.Context, terminate func()) | |||||
| } | } | ||||
| // RunAtShutdown creates a go-routine to run the provided function at shutdown | // RunAtShutdown creates a go-routine to run the provided function at shutdown | ||||
| func (g *gracefulManager) RunAtShutdown(ctx context.Context, shutdown func()) { | |||||
| func (g *Manager) RunAtShutdown(ctx context.Context, shutdown func()) { | |||||
| go func() { | go func() { | ||||
| select { | select { | ||||
| case <-g.IsShutdown(): | case <-g.IsShutdown(): | ||||
| @@ -132,7 +142,7 @@ func (g *gracefulManager) RunAtShutdown(ctx context.Context, shutdown func()) { | |||||
| } | } | ||||
| // RunAtHammer creates a go-routine to run the provided function at shutdown | // RunAtHammer creates a go-routine to run the provided function at shutdown | ||||
| func (g *gracefulManager) RunAtHammer(ctx context.Context, hammer func()) { | |||||
| func (g *Manager) RunAtHammer(ctx context.Context, hammer func()) { | |||||
| go func() { | go func() { | ||||
| select { | select { | ||||
| case <-g.IsHammer(): | case <-g.IsHammer(): | ||||
| @@ -141,7 +151,7 @@ func (g *gracefulManager) RunAtHammer(ctx context.Context, hammer func()) { | |||||
| } | } | ||||
| }() | }() | ||||
| } | } | ||||
| func (g *gracefulManager) doShutdown() { | |||||
| func (g *Manager) doShutdown() { | |||||
| if !g.setStateTransition(stateRunning, stateShuttingDown) { | if !g.setStateTransition(stateRunning, stateShuttingDown) { | ||||
| return | return | ||||
| } | } | ||||
| @@ -158,48 +168,47 @@ func (g *gracefulManager) doShutdown() { | |||||
| g.doHammerTime(0) | g.doHammerTime(0) | ||||
| <-time.After(1 * time.Second) | <-time.After(1 * time.Second) | ||||
| g.doTerminate() | g.doTerminate() | ||||
| g.WaitForTerminate() | |||||
| g.lock.Lock() | |||||
| close(g.done) | |||||
| g.lock.Unlock() | |||||
| }() | }() | ||||
| } | } | ||||
| func (g *gracefulManager) doHammerTime(d time.Duration) { | |||||
| func (g *Manager) doHammerTime(d time.Duration) { | |||||
| time.Sleep(d) | time.Sleep(d) | ||||
| g.lock.Lock() | |||||
| select { | select { | ||||
| case <-g.hammer: | case <-g.hammer: | ||||
| default: | default: | ||||
| log.Warn("Setting Hammer condition") | log.Warn("Setting Hammer condition") | ||||
| close(g.hammer) | close(g.hammer) | ||||
| } | } | ||||
| g.lock.Unlock() | |||||
| } | } | ||||
| func (g *gracefulManager) doTerminate() { | |||||
| func (g *Manager) doTerminate() { | |||||
| if !g.setStateTransition(stateShuttingDown, stateTerminate) { | if !g.setStateTransition(stateShuttingDown, stateTerminate) { | ||||
| return | return | ||||
| } | } | ||||
| g.lock.Lock() | g.lock.Lock() | ||||
| close(g.terminate) | |||||
| select { | |||||
| case <-g.terminate: | |||||
| default: | |||||
| log.Warn("Terminating") | |||||
| close(g.terminate) | |||||
| } | |||||
| g.lock.Unlock() | g.lock.Unlock() | ||||
| } | } | ||||
| // IsChild returns if the current process is a child of previous Gitea process | // IsChild returns if the current process is a child of previous Gitea process | ||||
| func (g *gracefulManager) IsChild() bool { | |||||
| func (g *Manager) IsChild() bool { | |||||
| return g.isChild | return g.isChild | ||||
| } | } | ||||
| // IsShutdown returns a channel which will be closed at shutdown. | // IsShutdown returns a channel which will be closed at shutdown. | ||||
| // The order of closure is IsShutdown, IsHammer (potentially), IsTerminate | // The order of closure is IsShutdown, IsHammer (potentially), IsTerminate | ||||
| func (g *gracefulManager) IsShutdown() <-chan struct{} { | |||||
| g.lock.RLock() | |||||
| if g.shutdown == nil { | |||||
| g.lock.RUnlock() | |||||
| g.lock.Lock() | |||||
| if g.shutdown == nil { | |||||
| g.shutdown = make(chan struct{}) | |||||
| } | |||||
| defer g.lock.Unlock() | |||||
| return g.shutdown | |||||
| } | |||||
| defer g.lock.RUnlock() | |||||
| func (g *Manager) IsShutdown() <-chan struct{} { | |||||
| return g.shutdown | return g.shutdown | ||||
| } | } | ||||
| @@ -207,65 +216,43 @@ func (g *gracefulManager) IsShutdown() <-chan struct{} { | |||||
| // The order of closure is IsShutdown, IsHammer (potentially), IsTerminate | // The order of closure is IsShutdown, IsHammer (potentially), IsTerminate | ||||
| // Servers running within the running server wait group should respond to IsHammer | // Servers running within the running server wait group should respond to IsHammer | ||||
| // if not shutdown already | // if not shutdown already | ||||
| func (g *gracefulManager) IsHammer() <-chan struct{} { | |||||
| g.lock.RLock() | |||||
| if g.hammer == nil { | |||||
| g.lock.RUnlock() | |||||
| g.lock.Lock() | |||||
| if g.hammer == nil { | |||||
| g.hammer = make(chan struct{}) | |||||
| } | |||||
| defer g.lock.Unlock() | |||||
| return g.hammer | |||||
| } | |||||
| defer g.lock.RUnlock() | |||||
| func (g *Manager) IsHammer() <-chan struct{} { | |||||
| return g.hammer | return g.hammer | ||||
| } | } | ||||
| // IsTerminate returns a channel which will be closed at terminate | // IsTerminate returns a channel which will be closed at terminate | ||||
| // The order of closure is IsShutdown, IsHammer (potentially), IsTerminate | // The order of closure is IsShutdown, IsHammer (potentially), IsTerminate | ||||
| // IsTerminate will only close once all running servers have stopped | // IsTerminate will only close once all running servers have stopped | ||||
| func (g *gracefulManager) IsTerminate() <-chan struct{} { | |||||
| g.lock.RLock() | |||||
| if g.terminate == nil { | |||||
| g.lock.RUnlock() | |||||
| g.lock.Lock() | |||||
| if g.terminate == nil { | |||||
| g.terminate = make(chan struct{}) | |||||
| } | |||||
| defer g.lock.Unlock() | |||||
| return g.terminate | |||||
| } | |||||
| defer g.lock.RUnlock() | |||||
| func (g *Manager) IsTerminate() <-chan struct{} { | |||||
| return g.terminate | return g.terminate | ||||
| } | } | ||||
| // ServerDone declares a running server done and subtracts one from the | // ServerDone declares a running server done and subtracts one from the | ||||
| // running server wait group. Users probably do not want to call this | // running server wait group. Users probably do not want to call this | ||||
| // and should use one of the RunWithShutdown* functions | // and should use one of the RunWithShutdown* functions | ||||
| func (g *gracefulManager) ServerDone() { | |||||
| func (g *Manager) ServerDone() { | |||||
| g.runningServerWaitGroup.Done() | g.runningServerWaitGroup.Done() | ||||
| } | } | ||||
| // WaitForServers waits for all running servers to finish. Users should probably | // WaitForServers waits for all running servers to finish. Users should probably | ||||
| // instead use AtTerminate or IsTerminate | // instead use AtTerminate or IsTerminate | ||||
| func (g *gracefulManager) WaitForServers() { | |||||
| func (g *Manager) WaitForServers() { | |||||
| g.runningServerWaitGroup.Wait() | g.runningServerWaitGroup.Wait() | ||||
| } | } | ||||
| // WaitForTerminate waits for all terminating actions to finish. | // WaitForTerminate waits for all terminating actions to finish. | ||||
| // Only the main go-routine should use this | // Only the main go-routine should use this | ||||
| func (g *gracefulManager) WaitForTerminate() { | |||||
| func (g *Manager) WaitForTerminate() { | |||||
| g.terminateWaitGroup.Wait() | g.terminateWaitGroup.Wait() | ||||
| } | } | ||||
| func (g *gracefulManager) getState() state { | |||||
| func (g *Manager) getState() state { | |||||
| g.lock.RLock() | g.lock.RLock() | ||||
| defer g.lock.RUnlock() | defer g.lock.RUnlock() | ||||
| return g.state | return g.state | ||||
| } | } | ||||
| func (g *gracefulManager) setStateTransition(old, new state) bool { | |||||
| func (g *Manager) setStateTransition(old, new state) bool { | |||||
| if old != g.getState() { | if old != g.getState() { | ||||
| return false | return false | ||||
| } | } | ||||
| @@ -279,7 +266,7 @@ func (g *gracefulManager) setStateTransition(old, new state) bool { | |||||
| return true | return true | ||||
| } | } | ||||
| func (g *gracefulManager) setState(st state) { | |||||
| func (g *Manager) setState(st state) { | |||||
| g.lock.Lock() | g.lock.Lock() | ||||
| defer g.lock.Unlock() | defer g.lock.Unlock() | ||||
| @@ -288,6 +275,31 @@ func (g *gracefulManager) setState(st state) { | |||||
| // InformCleanup tells the cleanup wait group that we have either taken a listener | // InformCleanup tells the cleanup wait group that we have either taken a listener | ||||
| // or will not be taking a listener | // or will not be taking a listener | ||||
| func (g *gracefulManager) InformCleanup() { | |||||
| func (g *Manager) InformCleanup() { | |||||
| g.createServerWaitGroup.Done() | g.createServerWaitGroup.Done() | ||||
| } | } | ||||
| // Done allows the manager to be viewed as a context.Context, it returns a channel that is closed when the server is finished terminating | |||||
| func (g *Manager) Done() <-chan struct{} { | |||||
| return g.done | |||||
| } | |||||
| // Err allows the manager to be viewed as a context.Context done at Terminate, it returns ErrTerminate | |||||
| func (g *Manager) Err() error { | |||||
| select { | |||||
| case <-g.Done(): | |||||
| return ErrTerminate | |||||
| default: | |||||
| return nil | |||||
| } | |||||
| } | |||||
| // Value allows the manager to be viewed as a context.Context done at Terminate, it has no values | |||||
| func (g *Manager) Value(key interface{}) interface{} { | |||||
| return nil | |||||
| } | |||||
| // Deadline returns nil as there is no fixed Deadline for the manager, it allows the manager to be viewed as a context.Context | |||||
| func (g *Manager) Deadline() (deadline time.Time, ok bool) { | |||||
| return | |||||
| } | |||||
| @@ -19,7 +19,8 @@ import ( | |||||
| "code.gitea.io/gitea/modules/setting" | "code.gitea.io/gitea/modules/setting" | ||||
| ) | ) | ||||
| type gracefulManager struct { | |||||
| // Manager manages the graceful shutdown process | |||||
| type Manager struct { | |||||
| isChild bool | isChild bool | ||||
| forked bool | forked bool | ||||
| lock *sync.RWMutex | lock *sync.RWMutex | ||||
| @@ -27,27 +28,37 @@ type gracefulManager struct { | |||||
| shutdown chan struct{} | shutdown chan struct{} | ||||
| hammer chan struct{} | hammer chan struct{} | ||||
| terminate chan struct{} | terminate chan struct{} | ||||
| done chan struct{} | |||||
| runningServerWaitGroup sync.WaitGroup | runningServerWaitGroup sync.WaitGroup | ||||
| createServerWaitGroup sync.WaitGroup | createServerWaitGroup sync.WaitGroup | ||||
| terminateWaitGroup sync.WaitGroup | terminateWaitGroup sync.WaitGroup | ||||
| } | } | ||||
| func newGracefulManager(ctx context.Context) *gracefulManager { | |||||
| manager := &gracefulManager{ | |||||
| func newGracefulManager(ctx context.Context) *Manager { | |||||
| manager := &Manager{ | |||||
| isChild: len(os.Getenv(listenFDs)) > 0 && os.Getppid() > 1, | isChild: len(os.Getenv(listenFDs)) > 0 && os.Getppid() > 1, | ||||
| lock: &sync.RWMutex{}, | lock: &sync.RWMutex{}, | ||||
| } | } | ||||
| manager.createServerWaitGroup.Add(numberOfServersToCreate) | manager.createServerWaitGroup.Add(numberOfServersToCreate) | ||||
| manager.Run(ctx) | |||||
| manager.start(ctx) | |||||
| return manager | return manager | ||||
| } | } | ||||
| func (g *gracefulManager) Run(ctx context.Context) { | |||||
| func (g *Manager) start(ctx context.Context) { | |||||
| // Make channels | |||||
| g.terminate = make(chan struct{}) | |||||
| g.shutdown = make(chan struct{}) | |||||
| g.hammer = make(chan struct{}) | |||||
| g.done = make(chan struct{}) | |||||
| // Set the running state & handle signals | |||||
| g.setState(stateRunning) | g.setState(stateRunning) | ||||
| go g.handleSignals(ctx) | go g.handleSignals(ctx) | ||||
| c := make(chan struct{}) | |||||
| // Handle clean up of unused provided listeners and delayed start-up | |||||
| startupDone := make(chan struct{}) | |||||
| go func() { | go func() { | ||||
| defer close(c) | |||||
| defer close(startupDone) | |||||
| // Wait till we're done getting all of the listeners and then close | // Wait till we're done getting all of the listeners and then close | ||||
| // the unused ones | // the unused ones | ||||
| g.createServerWaitGroup.Wait() | g.createServerWaitGroup.Wait() | ||||
| @@ -58,9 +69,19 @@ func (g *gracefulManager) Run(ctx context.Context) { | |||||
| if setting.StartupTimeout > 0 { | if setting.StartupTimeout > 0 { | ||||
| go func() { | go func() { | ||||
| select { | select { | ||||
| case <-c: | |||||
| case <-startupDone: | |||||
| return | return | ||||
| case <-g.IsShutdown(): | case <-g.IsShutdown(): | ||||
| func() { | |||||
| // When waitgroup counter goes negative it will panic - we don't care about this so we can just ignore it. | |||||
| defer func() { | |||||
| _ = recover() | |||||
| }() | |||||
| // Ensure that the createServerWaitGroup stops waiting | |||||
| for { | |||||
| g.createServerWaitGroup.Done() | |||||
| } | |||||
| }() | |||||
| return | return | ||||
| case <-time.After(setting.StartupTimeout): | case <-time.After(setting.StartupTimeout): | ||||
| log.Error("Startup took too long! Shutting down") | log.Error("Startup took too long! Shutting down") | ||||
| @@ -70,7 +91,7 @@ func (g *gracefulManager) Run(ctx context.Context) { | |||||
| } | } | ||||
| } | } | ||||
| func (g *gracefulManager) handleSignals(ctx context.Context) { | |||||
| func (g *Manager) handleSignals(ctx context.Context) { | |||||
| signalChannel := make(chan os.Signal, 1) | signalChannel := make(chan os.Signal, 1) | ||||
| signal.Notify( | signal.Notify( | ||||
| @@ -123,7 +144,7 @@ func (g *gracefulManager) handleSignals(ctx context.Context) { | |||||
| } | } | ||||
| } | } | ||||
| func (g *gracefulManager) doFork() error { | |||||
| func (g *Manager) doFork() error { | |||||
| g.lock.Lock() | g.lock.Lock() | ||||
| if g.forked { | if g.forked { | ||||
| g.lock.Unlock() | g.lock.Unlock() | ||||
| @@ -139,7 +160,9 @@ func (g *gracefulManager) doFork() error { | |||||
| return err | return err | ||||
| } | } | ||||
| func (g *gracefulManager) RegisterServer() { | |||||
| // RegisterServer registers the running of a listening server, in the case of unix this means that the parent process can now die. | |||||
| // Any call to RegisterServer must be matched by a call to ServerDone | |||||
| func (g *Manager) RegisterServer() { | |||||
| KillParent() | KillParent() | ||||
| g.runningServerWaitGroup.Add(1) | g.runningServerWaitGroup.Add(1) | ||||
| } | } | ||||
| @@ -30,7 +30,8 @@ const ( | |||||
| acceptHammerCode = svc.Accepted(hammerCode) | acceptHammerCode = svc.Accepted(hammerCode) | ||||
| ) | ) | ||||
| type gracefulManager struct { | |||||
| // Manager manages the graceful shutdown process | |||||
| type Manager struct { | |||||
| ctx context.Context | ctx context.Context | ||||
| isChild bool | isChild bool | ||||
| lock *sync.RWMutex | lock *sync.RWMutex | ||||
| @@ -38,27 +39,37 @@ type gracefulManager struct { | |||||
| shutdown chan struct{} | shutdown chan struct{} | ||||
| hammer chan struct{} | hammer chan struct{} | ||||
| terminate chan struct{} | terminate chan struct{} | ||||
| done chan struct{} | |||||
| runningServerWaitGroup sync.WaitGroup | runningServerWaitGroup sync.WaitGroup | ||||
| createServerWaitGroup sync.WaitGroup | createServerWaitGroup sync.WaitGroup | ||||
| terminateWaitGroup sync.WaitGroup | terminateWaitGroup sync.WaitGroup | ||||
| } | } | ||||
| func newGracefulManager(ctx context.Context) *gracefulManager { | |||||
| manager := &gracefulManager{ | |||||
| func newGracefulManager(ctx context.Context) *Manager { | |||||
| manager := &Manager{ | |||||
| isChild: false, | isChild: false, | ||||
| lock: &sync.RWMutex{}, | lock: &sync.RWMutex{}, | ||||
| ctx: ctx, | ctx: ctx, | ||||
| } | } | ||||
| manager.createServerWaitGroup.Add(numberOfServersToCreate) | manager.createServerWaitGroup.Add(numberOfServersToCreate) | ||||
| manager.Run() | |||||
| manager.start() | |||||
| return manager | return manager | ||||
| } | } | ||||
| func (g *gracefulManager) Run() { | |||||
| func (g *Manager) start() { | |||||
| // Make channels | |||||
| g.terminate = make(chan struct{}) | |||||
| g.shutdown = make(chan struct{}) | |||||
| g.hammer = make(chan struct{}) | |||||
| g.done = make(chan struct{}) | |||||
| // Set the running state | |||||
| g.setState(stateRunning) | g.setState(stateRunning) | ||||
| if skip, _ := strconv.ParseBool(os.Getenv("SKIP_MINWINSVC")); skip { | if skip, _ := strconv.ParseBool(os.Getenv("SKIP_MINWINSVC")); skip { | ||||
| return | return | ||||
| } | } | ||||
| // Make SVC process | |||||
| run := svc.Run | run := svc.Run | ||||
| isInteractive, err := svc.IsAnInteractiveSession() | isInteractive, err := svc.IsAnInteractiveSession() | ||||
| if err != nil { | if err != nil { | ||||
| @@ -71,8 +82,8 @@ func (g *gracefulManager) Run() { | |||||
| go run(WindowsServiceName, g) | go run(WindowsServiceName, g) | ||||
| } | } | ||||
| // Execute makes gracefulManager implement svc.Handler | |||||
| func (g *gracefulManager) Execute(args []string, changes <-chan svc.ChangeRequest, status chan<- svc.Status) (svcSpecificEC bool, exitCode uint32) { | |||||
| // Execute makes Manager implement svc.Handler | |||||
| func (g *Manager) Execute(args []string, changes <-chan svc.ChangeRequest, status chan<- svc.Status) (svcSpecificEC bool, exitCode uint32) { | |||||
| if setting.StartupTimeout > 0 { | if setting.StartupTimeout > 0 { | ||||
| status <- svc.Status{State: svc.StartPending} | status <- svc.Status{State: svc.StartPending} | ||||
| } else { | } else { | ||||
| @@ -141,11 +152,13 @@ hammerLoop: | |||||
| return false, 0 | return false, 0 | ||||
| } | } | ||||
| func (g *gracefulManager) RegisterServer() { | |||||
| // RegisterServer registers the running of a listening server. | |||||
| // Any call to RegisterServer must be matched by a call to ServerDone | |||||
| func (g *Manager) RegisterServer() { | |||||
| g.runningServerWaitGroup.Add(1) | g.runningServerWaitGroup.Add(1) | ||||
| } | } | ||||
| func (g *gracefulManager) awaitServer(limit time.Duration) bool { | |||||
| func (g *Manager) awaitServer(limit time.Duration) bool { | |||||
| c := make(chan struct{}) | c := make(chan struct{}) | ||||
| go func() { | go func() { | ||||
| defer close(c) | defer close(c) | ||||
| @@ -101,7 +101,7 @@ func CloseProvidedListeners() error { | |||||
| // creates a new one using net.Listen. | // creates a new one using net.Listen. | ||||
| func GetListener(network, address string) (net.Listener, error) { | func GetListener(network, address string) (net.Listener, error) { | ||||
| // Add a deferral to say that we've tried to grab a listener | // Add a deferral to say that we've tried to grab a listener | ||||
| defer Manager.InformCleanup() | |||||
| defer GetManager().InformCleanup() | |||||
| switch network { | switch network { | ||||
| case "tcp", "tcp4", "tcp6": | case "tcp", "tcp4", "tcp6": | ||||
| tcpAddr, err := net.ResolveTCPAddr(network, address) | tcpAddr, err := net.ResolveTCPAddr(network, address) | ||||
| @@ -22,7 +22,7 @@ var killParent sync.Once | |||||
| // KillParent sends the kill signal to the parent process if we are a child | // KillParent sends the kill signal to the parent process if we are a child | ||||
| func KillParent() { | func KillParent() { | ||||
| killParent.Do(func() { | killParent.Do(func() { | ||||
| if Manager.IsChild() { | |||||
| if GetManager().IsChild() { | |||||
| ppid := syscall.Getppid() | ppid := syscall.Getppid() | ||||
| if ppid > 1 { | if ppid > 1 { | ||||
| _ = syscall.Kill(ppid, syscall.SIGTERM) | _ = syscall.Kill(ppid, syscall.SIGTERM) | ||||
| @@ -47,7 +47,7 @@ type Server struct { | |||||
| // NewServer creates a server on network at provided address | // NewServer creates a server on network at provided address | ||||
| func NewServer(network, address string) *Server { | func NewServer(network, address string) *Server { | ||||
| if Manager.IsChild() { | |||||
| if GetManager().IsChild() { | |||||
| log.Info("Restarting new server: %s:%s on PID: %d", network, address, os.Getpid()) | log.Info("Restarting new server: %s:%s on PID: %d", network, address, os.Getpid()) | ||||
| } else { | } else { | ||||
| log.Info("Starting new server: %s:%s on PID: %d", network, address, os.Getpid()) | log.Info("Starting new server: %s:%s on PID: %d", network, address, os.Getpid()) | ||||
| @@ -138,12 +138,12 @@ func (srv *Server) ListenAndServeTLSConfig(tlsConfig *tls.Config, serve ServeFun | |||||
| func (srv *Server) Serve(serve ServeFunction) error { | func (srv *Server) Serve(serve ServeFunction) error { | ||||
| defer log.Debug("Serve() returning... (PID: %d)", syscall.Getpid()) | defer log.Debug("Serve() returning... (PID: %d)", syscall.Getpid()) | ||||
| srv.setState(stateRunning) | srv.setState(stateRunning) | ||||
| Manager.RegisterServer() | |||||
| GetManager().RegisterServer() | |||||
| err := serve(srv.listener) | err := serve(srv.listener) | ||||
| log.Debug("Waiting for connections to finish... (PID: %d)", syscall.Getpid()) | log.Debug("Waiting for connections to finish... (PID: %d)", syscall.Getpid()) | ||||
| srv.wg.Wait() | srv.wg.Wait() | ||||
| srv.setState(stateTerminate) | srv.setState(stateTerminate) | ||||
| Manager.ServerDone() | |||||
| GetManager().ServerDone() | |||||
| // use of closed means that the listeners are closed - i.e. we should be shutting down - return nil | // use of closed means that the listeners are closed - i.e. we should be shutting down - return nil | ||||
| if err != nil && strings.Contains(err.Error(), "use of closed") { | if err != nil && strings.Contains(err.Error(), "use of closed") { | ||||
| return nil | return nil | ||||
| @@ -14,15 +14,15 @@ import ( | |||||
| // awaitShutdown waits for the shutdown signal from the Manager | // awaitShutdown waits for the shutdown signal from the Manager | ||||
| func (srv *Server) awaitShutdown() { | func (srv *Server) awaitShutdown() { | ||||
| select { | select { | ||||
| case <-Manager.IsShutdown(): | |||||
| case <-GetManager().IsShutdown(): | |||||
| // Shutdown | // Shutdown | ||||
| srv.doShutdown() | srv.doShutdown() | ||||
| case <-Manager.IsHammer(): | |||||
| case <-GetManager().IsHammer(): | |||||
| // Hammer | // Hammer | ||||
| srv.doShutdown() | srv.doShutdown() | ||||
| srv.doHammer() | srv.doHammer() | ||||
| } | } | ||||
| <-Manager.IsHammer() | |||||
| <-GetManager().IsHammer() | |||||
| srv.doHammer() | srv.doHammer() | ||||
| } | } | ||||
| @@ -6,6 +6,7 @@ package code | |||||
| import ( | import ( | ||||
| "fmt" | "fmt" | ||||
| "os" | |||||
| "strconv" | "strconv" | ||||
| "strings" | "strings" | ||||
| "time" | "time" | ||||
| @@ -34,10 +35,11 @@ func InitRepoIndexer() { | |||||
| return | return | ||||
| } | } | ||||
| waitChannel := make(chan time.Duration) | waitChannel := make(chan time.Duration) | ||||
| // FIXME: graceful: This should use a persistable queue | |||||
| repoIndexerOperationQueue = make(chan repoIndexerOperation, setting.Indexer.UpdateQueueLength) | repoIndexerOperationQueue = make(chan repoIndexerOperation, setting.Indexer.UpdateQueueLength) | ||||
| go func() { | go func() { | ||||
| start := time.Now() | start := time.Now() | ||||
| log.Info("Initializing Repository Indexer") | |||||
| log.Info("PID: %d: Initializing Repository Indexer", os.Getpid()) | |||||
| initRepoIndexer(populateRepoIndexerAsynchronously) | initRepoIndexer(populateRepoIndexerAsynchronously) | ||||
| go processRepoIndexerOperationQueue() | go processRepoIndexerOperationQueue() | ||||
| waitChannel <- time.Since(start) | waitChannel <- time.Since(start) | ||||
| @@ -45,7 +47,7 @@ func InitRepoIndexer() { | |||||
| if setting.Indexer.StartupTimeout > 0 { | if setting.Indexer.StartupTimeout > 0 { | ||||
| go func() { | go func() { | ||||
| timeout := setting.Indexer.StartupTimeout | timeout := setting.Indexer.StartupTimeout | ||||
| if graceful.Manager.IsChild() && setting.GracefulHammerTime > 0 { | |||||
| if graceful.GetManager().IsChild() && setting.GracefulHammerTime > 0 { | |||||
| timeout += setting.GracefulHammerTime | timeout += setting.GracefulHammerTime | ||||
| } | } | ||||
| select { | select { | ||||
| @@ -70,13 +72,6 @@ func populateRepoIndexerAsynchronously() error { | |||||
| return nil | return nil | ||||
| } | } | ||||
| // if there is any existing repo indexer metadata in the DB, delete it | |||||
| // since we are starting afresh. Also, xorm requires deletes to have a | |||||
| // condition, and we want to delete everything, thus 1=1. | |||||
| if err := models.DeleteAllRecords("repo_indexer_status"); err != nil { | |||||
| return err | |||||
| } | |||||
| var maxRepoID int64 | var maxRepoID int64 | ||||
| if maxRepoID, err = models.GetMaxID("repository"); err != nil { | if maxRepoID, err = models.GetMaxID("repository"); err != nil { | ||||
| return err | return err | ||||
| @@ -87,44 +82,59 @@ func populateRepoIndexerAsynchronously() error { | |||||
| // populateRepoIndexer populate the repo indexer with pre-existing data. This | // populateRepoIndexer populate the repo indexer with pre-existing data. This | ||||
| // should only be run when the indexer is created for the first time. | // should only be run when the indexer is created for the first time. | ||||
| // FIXME: graceful: This should use a persistable queue | |||||
| func populateRepoIndexer(maxRepoID int64) { | func populateRepoIndexer(maxRepoID int64) { | ||||
| log.Info("Populating the repo indexer with existing repositories") | log.Info("Populating the repo indexer with existing repositories") | ||||
| isShutdown := graceful.GetManager().IsShutdown() | |||||
| // start with the maximum existing repo ID and work backwards, so that we | // start with the maximum existing repo ID and work backwards, so that we | ||||
| // don't include repos that are created after gitea starts; such repos will | // don't include repos that are created after gitea starts; such repos will | ||||
| // already be added to the indexer, and we don't need to add them again. | // already be added to the indexer, and we don't need to add them again. | ||||
| for maxRepoID > 0 { | for maxRepoID > 0 { | ||||
| repos := make([]*models.Repository, 0, models.RepositoryListDefaultPageSize) | |||||
| err := models.FindByMaxID(maxRepoID, models.RepositoryListDefaultPageSize, &repos) | |||||
| select { | |||||
| case <-isShutdown: | |||||
| log.Info("Repository Indexer population shutdown before completion") | |||||
| return | |||||
| default: | |||||
| } | |||||
| ids, err := models.GetUnindexedRepos(maxRepoID, 0, 50) | |||||
| if err != nil { | if err != nil { | ||||
| log.Error("populateRepoIndexer: %v", err) | log.Error("populateRepoIndexer: %v", err) | ||||
| return | return | ||||
| } else if len(repos) == 0 { | |||||
| } else if len(ids) == 0 { | |||||
| break | break | ||||
| } | } | ||||
| for _, repo := range repos { | |||||
| for _, id := range ids { | |||||
| select { | |||||
| case <-isShutdown: | |||||
| log.Info("Repository Indexer population shutdown before completion") | |||||
| return | |||||
| default: | |||||
| } | |||||
| repoIndexerOperationQueue <- repoIndexerOperation{ | repoIndexerOperationQueue <- repoIndexerOperation{ | ||||
| repoID: repo.ID, | |||||
| repoID: id, | |||||
| deleted: false, | deleted: false, | ||||
| } | } | ||||
| maxRepoID = repo.ID - 1 | |||||
| maxRepoID = id - 1 | |||||
| } | } | ||||
| } | } | ||||
| log.Info("Done populating the repo indexer with existing repositories") | |||||
| log.Info("Done (re)populating the repo indexer with existing repositories") | |||||
| } | } | ||||
| func updateRepoIndexer(repoID int64) error { | func updateRepoIndexer(repoID int64) error { | ||||
| repo, err := models.GetRepositoryByID(repoID) | repo, err := models.GetRepositoryByID(repoID) | ||||
| if err != nil { | if err != nil { | ||||
| return err | |||||
| return fmt.Errorf("UpdateRepoIndexer: Unable to GetRepositoryByID: %d, Error: %v", repoID, err) | |||||
| } | } | ||||
| sha, err := getDefaultBranchSha(repo) | sha, err := getDefaultBranchSha(repo) | ||||
| if err != nil { | if err != nil { | ||||
| return err | |||||
| return fmt.Errorf("UpdateRepoIndexer: Unable to GetDefaultBranchSha for: %s/%s, Error: %v", repo.MustOwnerName(), repo.Name, err) | |||||
| } | } | ||||
| changes, err := getRepoChanges(repo, sha) | changes, err := getRepoChanges(repo, sha) | ||||
| if err != nil { | if err != nil { | ||||
| return err | |||||
| return fmt.Errorf("UpdateRepoIndexer: Unable to GetRepoChanges for: %s/%s Sha: %s Error: %v", repo.MustOwnerName(), repo.Name, sha, err) | |||||
| } else if changes == nil { | } else if changes == nil { | ||||
| return nil | return nil | ||||
| } | } | ||||
| @@ -132,16 +142,16 @@ func updateRepoIndexer(repoID int64) error { | |||||
| batch := RepoIndexerBatch() | batch := RepoIndexerBatch() | ||||
| for _, update := range changes.Updates { | for _, update := range changes.Updates { | ||||
| if err := addUpdate(update, repo, batch); err != nil { | if err := addUpdate(update, repo, batch); err != nil { | ||||
| return err | |||||
| return fmt.Errorf("UpdateRepoIndexer: Unable to addUpdate to: %s/%s Sha: %s, update: %s(%s) Error: %v", repo.MustOwnerName(), repo.Name, sha, update.Filename, update.BlobSha, err) | |||||
| } | } | ||||
| } | } | ||||
| for _, filename := range changes.RemovedFilenames { | for _, filename := range changes.RemovedFilenames { | ||||
| if err := addDelete(filename, repo, batch); err != nil { | if err := addDelete(filename, repo, batch); err != nil { | ||||
| return err | |||||
| return fmt.Errorf("UpdateRepoIndexer: Unable to addDelete to: %s/%s Sha: %s, filename: %s Error: %v", repo.MustOwnerName(), repo.Name, sha, filename, err) | |||||
| } | } | ||||
| } | } | ||||
| if err = batch.Flush(); err != nil { | if err = batch.Flush(); err != nil { | ||||
| return err | |||||
| return fmt.Errorf("UpdateRepoIndexer: Unable to flush batch to indexer for repo: %s/%s Error: %v", repo.MustOwnerName(), repo.Name, err) | |||||
| } | } | ||||
| return repo.UpdateIndexerStatus(sha) | return repo.UpdateIndexerStatus(sha) | ||||
| } | } | ||||
| @@ -322,20 +332,26 @@ func nonGenesisChanges(repo *models.Repository, revision string) (*repoChanges, | |||||
| func processRepoIndexerOperationQueue() { | func processRepoIndexerOperationQueue() { | ||||
| for { | for { | ||||
| op := <-repoIndexerOperationQueue | |||||
| var err error | |||||
| if op.deleted { | |||||
| if err = deleteRepoFromIndexer(op.repoID); err != nil { | |||||
| log.Error("deleteRepoFromIndexer: %v", err) | |||||
| select { | |||||
| case op := <-repoIndexerOperationQueue: | |||||
| var err error | |||||
| if op.deleted { | |||||
| if err = deleteRepoFromIndexer(op.repoID); err != nil { | |||||
| log.Error("DeleteRepoFromIndexer: %v", err) | |||||
| } | |||||
| } else { | |||||
| if err = updateRepoIndexer(op.repoID); err != nil { | |||||
| log.Error("updateRepoIndexer: %v", err) | |||||
| } | |||||
| } | } | ||||
| } else { | |||||
| if err = updateRepoIndexer(op.repoID); err != nil { | |||||
| log.Error("updateRepoIndexer: %v", err) | |||||
| for _, watcher := range op.watchers { | |||||
| watcher <- err | |||||
| } | } | ||||
| case <-graceful.GetManager().IsShutdown(): | |||||
| log.Info("PID: %d Repository indexer queue processing stopped", os.Getpid()) | |||||
| return | |||||
| } | } | ||||
| for _, watcher := range op.watchers { | |||||
| watcher <- err | |||||
| } | |||||
| } | } | ||||
| } | } | ||||
| @@ -5,9 +5,13 @@ | |||||
| package code | package code | ||||
| import ( | import ( | ||||
| "context" | |||||
| "os" | |||||
| "strings" | "strings" | ||||
| "sync" | "sync" | ||||
| "code.gitea.io/gitea/models" | |||||
| "code.gitea.io/gitea/modules/graceful" | |||||
| "code.gitea.io/gitea/modules/log" | "code.gitea.io/gitea/modules/log" | ||||
| "code.gitea.io/gitea/modules/setting" | "code.gitea.io/gitea/modules/setting" | ||||
| @@ -104,21 +108,50 @@ func (update RepoIndexerUpdate) AddToFlushingBatch(batch rupture.FlushingBatch) | |||||
| func initRepoIndexer(populateIndexer func() error) { | func initRepoIndexer(populateIndexer func() error) { | ||||
| indexer, err := openIndexer(setting.Indexer.RepoPath, repoIndexerLatestVersion) | indexer, err := openIndexer(setting.Indexer.RepoPath, repoIndexerLatestVersion) | ||||
| if err != nil { | if err != nil { | ||||
| log.Fatal("InitRepoIndexer: %v", err) | |||||
| log.Fatal("InitRepoIndexer %s: %v", setting.Indexer.RepoPath, err) | |||||
| } | } | ||||
| if indexer != nil { | if indexer != nil { | ||||
| indexerHolder.set(indexer) | indexerHolder.set(indexer) | ||||
| closeAtTerminate() | |||||
| // Continue population from where left off | |||||
| if err = populateIndexer(); err != nil { | |||||
| log.Fatal("PopulateRepoIndex: %v", err) | |||||
| } | |||||
| return | return | ||||
| } | } | ||||
| if err = createRepoIndexer(setting.Indexer.RepoPath, repoIndexerLatestVersion); err != nil { | if err = createRepoIndexer(setting.Indexer.RepoPath, repoIndexerLatestVersion); err != nil { | ||||
| log.Fatal("CreateRepoIndexer: %v", err) | log.Fatal("CreateRepoIndexer: %v", err) | ||||
| } | } | ||||
| closeAtTerminate() | |||||
| // if there is any existing repo indexer metadata in the DB, delete it | |||||
| // since we are starting afresh. Also, xorm requires deletes to have a | |||||
| // condition, and we want to delete everything, thus 1=1. | |||||
| if err := models.DeleteAllRecords("repo_indexer_status"); err != nil { | |||||
| log.Fatal("DeleteAllRepoIndexerStatus: %v", err) | |||||
| } | |||||
| if err = populateIndexer(); err != nil { | if err = populateIndexer(); err != nil { | ||||
| log.Fatal("PopulateRepoIndex: %v", err) | log.Fatal("PopulateRepoIndex: %v", err) | ||||
| } | } | ||||
| } | } | ||||
| func closeAtTerminate() { | |||||
| graceful.GetManager().RunAtTerminate(context.Background(), func() { | |||||
| log.Debug("Closing repo indexer") | |||||
| indexer := indexerHolder.get() | |||||
| if indexer != nil { | |||||
| err := indexer.Close() | |||||
| if err != nil { | |||||
| log.Error("Error whilst closing the repository indexer: %v", err) | |||||
| } | |||||
| } | |||||
| log.Info("PID: %d Repository Indexer closed", os.Getpid()) | |||||
| }) | |||||
| } | |||||
| // createRepoIndexer create a repo indexer if one does not already exist | // createRepoIndexer create a repo indexer if one does not already exist | ||||
| func createRepoIndexer(path string, latestVersion int) error { | func createRepoIndexer(path string, latestVersion int) error { | ||||
| docMapping := bleve.NewDocumentMapping() | docMapping := bleve.NewDocumentMapping() | ||||
| @@ -172,7 +172,7 @@ func InitIssueIndexer(syncReindex bool) { | |||||
| } else if setting.Indexer.StartupTimeout > 0 { | } else if setting.Indexer.StartupTimeout > 0 { | ||||
| go func() { | go func() { | ||||
| timeout := setting.Indexer.StartupTimeout | timeout := setting.Indexer.StartupTimeout | ||||
| if graceful.Manager.IsChild() && setting.GracefulHammerTime > 0 { | |||||
| if graceful.GetManager().IsChild() && setting.GracefulHammerTime > 0 { | |||||
| timeout += setting.GracefulHammerTime | timeout += setting.GracefulHammerTime | ||||
| } | } | ||||
| select { | select { | ||||
| @@ -5,21 +5,28 @@ | |||||
| package migrations | package migrations | ||||
| import ( | import ( | ||||
| "context" | |||||
| "code.gitea.io/gitea/models" | "code.gitea.io/gitea/models" | ||||
| "code.gitea.io/gitea/modules/log" | "code.gitea.io/gitea/modules/log" | ||||
| "code.gitea.io/gitea/modules/structs" | "code.gitea.io/gitea/modules/structs" | ||||
| ) | ) | ||||
| // UpdateMigrationPosterID updates all migrated repositories' issues and comments posterID | // UpdateMigrationPosterID updates all migrated repositories' issues and comments posterID | ||||
| func UpdateMigrationPosterID() { | |||||
| func UpdateMigrationPosterID(ctx context.Context) { | |||||
| for _, gitService := range structs.SupportedFullGitService { | for _, gitService := range structs.SupportedFullGitService { | ||||
| if err := updateMigrationPosterIDByGitService(gitService); err != nil { | |||||
| select { | |||||
| case <-ctx.Done(): | |||||
| log.Warn("UpdateMigrationPosterID aborted due to shutdown before %s", gitService.Name()) | |||||
| default: | |||||
| } | |||||
| if err := updateMigrationPosterIDByGitService(ctx, gitService); err != nil { | |||||
| log.Error("updateMigrationPosterIDByGitService failed: %v", err) | log.Error("updateMigrationPosterIDByGitService failed: %v", err) | ||||
| } | } | ||||
| } | } | ||||
| } | } | ||||
| func updateMigrationPosterIDByGitService(tp structs.GitServiceType) error { | |||||
| func updateMigrationPosterIDByGitService(ctx context.Context, tp structs.GitServiceType) error { | |||||
| provider := tp.Name() | provider := tp.Name() | ||||
| if len(provider) == 0 { | if len(provider) == 0 { | ||||
| return nil | return nil | ||||
| @@ -28,6 +35,13 @@ func updateMigrationPosterIDByGitService(tp structs.GitServiceType) error { | |||||
| const batchSize = 100 | const batchSize = 100 | ||||
| var start int | var start int | ||||
| for { | for { | ||||
| select { | |||||
| case <-ctx.Done(): | |||||
| log.Warn("UpdateMigrationPosterIDByGitService(%s) aborted due to shutdown", tp.Name()) | |||||
| return nil | |||||
| default: | |||||
| } | |||||
| users, err := models.FindExternalUsersByProvider(models.FindExternalUserOptions{ | users, err := models.FindExternalUsersByProvider(models.FindExternalUserOptions{ | ||||
| Provider: provider, | Provider: provider, | ||||
| Start: start, | Start: start, | ||||
| @@ -38,6 +52,12 @@ func updateMigrationPosterIDByGitService(tp structs.GitServiceType) error { | |||||
| } | } | ||||
| for _, user := range users { | for _, user := range users { | ||||
| select { | |||||
| case <-ctx.Done(): | |||||
| log.Warn("UpdateMigrationPosterIDByGitService(%s) aborted due to shutdown", tp.Name()) | |||||
| return nil | |||||
| default: | |||||
| } | |||||
| externalUserID := user.ExternalID | externalUserID := user.ExternalID | ||||
| if err := models.UpdateMigrationsByType(tp, externalUserID, user.UserID); err != nil { | if err := models.UpdateMigrationsByType(tp, externalUserID, user.UserID); err != nil { | ||||
| log.Error("UpdateMigrationsByType type %s external user id %v to local user id %v failed: %v", tp.Name(), user.ExternalID, user.UserID, err) | log.Error("UpdateMigrationsByType type %s external user id %v to local user id %v failed: %v", tp.Name(), user.ExternalID, user.UserID, err) | ||||
| @@ -24,5 +24,5 @@ func listen(server *ssh.Server) { | |||||
| // Unused informs our cleanup routine that we will not be using a ssh port | // Unused informs our cleanup routine that we will not be using a ssh port | ||||
| func Unused() { | func Unused() { | ||||
| graceful.Manager.InformCleanup() | |||||
| graceful.GetManager().InformCleanup() | |||||
| } | } | ||||
| @@ -1,4 +1,5 @@ | |||||
| // Copyright 2016 The Gogs Authors. All rights reserved. | // Copyright 2016 The Gogs Authors. All rights reserved. | ||||
| // Copyright 2019 The Gitea Authors. All rights reserved. | |||||
| // Use of this source code is governed by a MIT-style | // Use of this source code is governed by a MIT-style | ||||
| // license that can be found in the LICENSE file. | // license that can be found in the LICENSE file. | ||||
| @@ -15,8 +16,9 @@ import ( | |||||
| // This queue is particularly useful for preventing duplicated task | // This queue is particularly useful for preventing duplicated task | ||||
| // of same purpose. | // of same purpose. | ||||
| type UniqueQueue struct { | type UniqueQueue struct { | ||||
| table *StatusTable | |||||
| queue chan string | |||||
| table *StatusTable | |||||
| queue chan string | |||||
| closed chan struct{} | |||||
| } | } | ||||
| // NewUniqueQueue initializes and returns a new UniqueQueue object. | // NewUniqueQueue initializes and returns a new UniqueQueue object. | ||||
| @@ -26,11 +28,43 @@ func NewUniqueQueue(queueLength int) *UniqueQueue { | |||||
| } | } | ||||
| return &UniqueQueue{ | return &UniqueQueue{ | ||||
| table: NewStatusTable(), | |||||
| queue: make(chan string, queueLength), | |||||
| table: NewStatusTable(), | |||||
| queue: make(chan string, queueLength), | |||||
| closed: make(chan struct{}), | |||||
| } | } | ||||
| } | } | ||||
| // Close closes this queue | |||||
| func (q *UniqueQueue) Close() { | |||||
| select { | |||||
| case <-q.closed: | |||||
| default: | |||||
| q.table.lock.Lock() | |||||
| select { | |||||
| case <-q.closed: | |||||
| default: | |||||
| close(q.closed) | |||||
| } | |||||
| q.table.lock.Unlock() | |||||
| } | |||||
| } | |||||
| // IsClosed returns a channel that is closed when this Queue is closed | |||||
| func (q *UniqueQueue) IsClosed() <-chan struct{} { | |||||
| return q.closed | |||||
| } | |||||
| // IDs returns the current ids in the pool | |||||
| func (q *UniqueQueue) IDs() []interface{} { | |||||
| q.table.lock.Lock() | |||||
| defer q.table.lock.Unlock() | |||||
| ids := make([]interface{}, 0, len(q.table.pool)) | |||||
| for id := range q.table.pool { | |||||
| ids = append(ids, id) | |||||
| } | |||||
| return ids | |||||
| } | |||||
| // Queue returns channel of queue for retrieving instances. | // Queue returns channel of queue for retrieving instances. | ||||
| func (q *UniqueQueue) Queue() <-chan string { | func (q *UniqueQueue) Queue() <-chan string { | ||||
| return q.queue | return q.queue | ||||
| @@ -45,18 +79,22 @@ func (q *UniqueQueue) Exist(id interface{}) bool { | |||||
| // AddFunc adds new instance to the queue with a custom runnable function, | // AddFunc adds new instance to the queue with a custom runnable function, | ||||
| // the queue is blocked until the function exits. | // the queue is blocked until the function exits. | ||||
| func (q *UniqueQueue) AddFunc(id interface{}, fn func()) { | func (q *UniqueQueue) AddFunc(id interface{}, fn func()) { | ||||
| if q.Exist(id) { | |||||
| return | |||||
| } | |||||
| idStr := com.ToStr(id) | idStr := com.ToStr(id) | ||||
| q.table.lock.Lock() | q.table.lock.Lock() | ||||
| if _, ok := q.table.pool[idStr]; ok { | |||||
| return | |||||
| } | |||||
| q.table.pool[idStr] = struct{}{} | q.table.pool[idStr] = struct{}{} | ||||
| if fn != nil { | if fn != nil { | ||||
| fn() | fn() | ||||
| } | } | ||||
| q.table.lock.Unlock() | q.table.lock.Unlock() | ||||
| q.queue <- idStr | |||||
| select { | |||||
| case <-q.closed: | |||||
| return | |||||
| case q.queue <- idStr: | |||||
| return | |||||
| } | |||||
| } | } | ||||
| // Add adds new instance to the queue. | // Add adds new instance to the queue. | ||||
| @@ -5,6 +5,7 @@ | |||||
| package webhook | package webhook | ||||
| import ( | import ( | ||||
| "context" | |||||
| "crypto/tls" | "crypto/tls" | ||||
| "fmt" | "fmt" | ||||
| "io/ioutil" | "io/ioutil" | ||||
| @@ -16,6 +17,7 @@ import ( | |||||
| "time" | "time" | ||||
| "code.gitea.io/gitea/models" | "code.gitea.io/gitea/models" | ||||
| "code.gitea.io/gitea/modules/graceful" | |||||
| "code.gitea.io/gitea/modules/log" | "code.gitea.io/gitea/modules/log" | ||||
| "code.gitea.io/gitea/modules/setting" | "code.gitea.io/gitea/modules/setting" | ||||
| "github.com/gobwas/glob" | "github.com/gobwas/glob" | ||||
| @@ -145,8 +147,14 @@ func Deliver(t *models.HookTask) error { | |||||
| } | } | ||||
| // DeliverHooks checks and delivers undelivered hooks. | // DeliverHooks checks and delivers undelivered hooks. | ||||
| // TODO: shoot more hooks at same time. | |||||
| func DeliverHooks() { | |||||
| // FIXME: graceful: This would likely benefit from either a worker pool with dummy queue | |||||
| // or a full queue. Then more hooks could be sent at same time. | |||||
| func DeliverHooks(ctx context.Context) { | |||||
| select { | |||||
| case <-ctx.Done(): | |||||
| return | |||||
| default: | |||||
| } | |||||
| tasks, err := models.FindUndeliveredHookTasks() | tasks, err := models.FindUndeliveredHookTasks() | ||||
| if err != nil { | if err != nil { | ||||
| log.Error("DeliverHooks: %v", err) | log.Error("DeliverHooks: %v", err) | ||||
| @@ -155,33 +163,50 @@ func DeliverHooks() { | |||||
| // Update hook task status. | // Update hook task status. | ||||
| for _, t := range tasks { | for _, t := range tasks { | ||||
| select { | |||||
| case <-ctx.Done(): | |||||
| return | |||||
| default: | |||||
| } | |||||
| if err = Deliver(t); err != nil { | if err = Deliver(t); err != nil { | ||||
| log.Error("deliver: %v", err) | log.Error("deliver: %v", err) | ||||
| } | } | ||||
| } | } | ||||
| // Start listening on new hook requests. | // Start listening on new hook requests. | ||||
| for repoIDStr := range hookQueue.Queue() { | |||||
| log.Trace("DeliverHooks [repo_id: %v]", repoIDStr) | |||||
| hookQueue.Remove(repoIDStr) | |||||
| for { | |||||
| select { | |||||
| case <-ctx.Done(): | |||||
| hookQueue.Close() | |||||
| return | |||||
| case repoIDStr := <-hookQueue.Queue(): | |||||
| log.Trace("DeliverHooks [repo_id: %v]", repoIDStr) | |||||
| hookQueue.Remove(repoIDStr) | |||||
| repoID, err := com.StrTo(repoIDStr).Int64() | |||||
| if err != nil { | |||||
| log.Error("Invalid repo ID: %s", repoIDStr) | |||||
| continue | |||||
| } | |||||
| repoID, err := com.StrTo(repoIDStr).Int64() | |||||
| if err != nil { | |||||
| log.Error("Invalid repo ID: %s", repoIDStr) | |||||
| continue | |||||
| } | |||||
| tasks, err := models.FindRepoUndeliveredHookTasks(repoID) | |||||
| if err != nil { | |||||
| log.Error("Get repository [%d] hook tasks: %v", repoID, err) | |||||
| continue | |||||
| } | |||||
| for _, t := range tasks { | |||||
| if err = Deliver(t); err != nil { | |||||
| log.Error("deliver: %v", err) | |||||
| tasks, err := models.FindRepoUndeliveredHookTasks(repoID) | |||||
| if err != nil { | |||||
| log.Error("Get repository [%d] hook tasks: %v", repoID, err) | |||||
| continue | |||||
| } | |||||
| for _, t := range tasks { | |||||
| select { | |||||
| case <-ctx.Done(): | |||||
| return | |||||
| default: | |||||
| } | |||||
| if err = Deliver(t); err != nil { | |||||
| log.Error("deliver: %v", err) | |||||
| } | |||||
| } | } | ||||
| } | } | ||||
| } | } | ||||
| } | } | ||||
| var ( | var ( | ||||
| @@ -234,5 +259,5 @@ func InitDeliverHooks() { | |||||
| }, | }, | ||||
| } | } | ||||
| go DeliverHooks() | |||||
| go graceful.GetManager().RunWithShutdownContext(DeliverHooks) | |||||
| } | } | ||||
| @@ -19,6 +19,7 @@ import ( | |||||
| "code.gitea.io/gitea/modules/context" | "code.gitea.io/gitea/modules/context" | ||||
| "code.gitea.io/gitea/modules/cron" | "code.gitea.io/gitea/modules/cron" | ||||
| "code.gitea.io/gitea/modules/git" | "code.gitea.io/gitea/modules/git" | ||||
| "code.gitea.io/gitea/modules/graceful" | |||||
| "code.gitea.io/gitea/modules/log" | "code.gitea.io/gitea/modules/log" | ||||
| "code.gitea.io/gitea/modules/process" | "code.gitea.io/gitea/modules/process" | ||||
| "code.gitea.io/gitea/modules/setting" | "code.gitea.io/gitea/modules/setting" | ||||
| @@ -171,10 +172,10 @@ func Dashboard(ctx *context.Context) { | |||||
| err = models.ReinitMissingRepositories() | err = models.ReinitMissingRepositories() | ||||
| case syncExternalUsers: | case syncExternalUsers: | ||||
| success = ctx.Tr("admin.dashboard.sync_external_users_started") | success = ctx.Tr("admin.dashboard.sync_external_users_started") | ||||
| go models.SyncExternalUsers() | |||||
| go graceful.GetManager().RunWithShutdownContext(models.SyncExternalUsers) | |||||
| case gitFsck: | case gitFsck: | ||||
| success = ctx.Tr("admin.dashboard.git_fsck_started") | success = ctx.Tr("admin.dashboard.git_fsck_started") | ||||
| go models.GitFsck() | |||||
| go graceful.GetManager().RunWithShutdownContext(models.GitFsck) | |||||
| case deleteGeneratedRepositoryAvatars: | case deleteGeneratedRepositoryAvatars: | ||||
| success = ctx.Tr("admin.dashboard.delete_generated_repository_avatars_success") | success = ctx.Tr("admin.dashboard.delete_generated_repository_avatars_success") | ||||
| err = models.RemoveRandomAvatars() | err = models.RemoveRandomAvatars() | ||||
| @@ -5,6 +5,7 @@ | |||||
| package routers | package routers | ||||
| import ( | import ( | ||||
| "context" | |||||
| "strings" | "strings" | ||||
| "time" | "time" | ||||
| @@ -53,11 +54,11 @@ func NewServices() { | |||||
| } | } | ||||
| // In case of problems connecting to DB, retry connection. Eg, PGSQL in Docker Container on Synology | // In case of problems connecting to DB, retry connection. Eg, PGSQL in Docker Container on Synology | ||||
| func initDBEngine() (err error) { | |||||
| func initDBEngine(ctx context.Context) (err error) { | |||||
| log.Info("Beginning ORM engine initialization.") | log.Info("Beginning ORM engine initialization.") | ||||
| for i := 0; i < setting.Database.DBConnectRetries; i++ { | for i := 0; i < setting.Database.DBConnectRetries; i++ { | ||||
| log.Info("ORM engine initialization attempt #%d/%d...", i+1, setting.Database.DBConnectRetries) | log.Info("ORM engine initialization attempt #%d/%d...", i+1, setting.Database.DBConnectRetries) | ||||
| if err = models.NewEngine(migrations.Migrate); err == nil { | |||||
| if err = models.NewEngine(ctx, migrations.Migrate); err == nil { | |||||
| break | break | ||||
| } else if i == setting.Database.DBConnectRetries-1 { | } else if i == setting.Database.DBConnectRetries-1 { | ||||
| return err | return err | ||||
| @@ -71,9 +72,9 @@ func initDBEngine() (err error) { | |||||
| } | } | ||||
| // GlobalInit is for global configuration reload-able. | // GlobalInit is for global configuration reload-able. | ||||
| func GlobalInit() { | |||||
| func GlobalInit(ctx context.Context) { | |||||
| setting.NewContext() | setting.NewContext() | ||||
| if err := git.Init(); err != nil { | |||||
| if err := git.Init(ctx); err != nil { | |||||
| log.Fatal("Git module init failed: %v", err) | log.Fatal("Git module init failed: %v", err) | ||||
| } | } | ||||
| setting.CheckLFSVersion() | setting.CheckLFSVersion() | ||||
| @@ -88,7 +89,7 @@ func GlobalInit() { | |||||
| highlight.NewContext() | highlight.NewContext() | ||||
| external.RegisterParsers() | external.RegisterParsers() | ||||
| markup.Init() | markup.Init() | ||||
| if err := initDBEngine(); err == nil { | |||||
| if err := initDBEngine(ctx); err == nil { | |||||
| log.Info("ORM engine initialization successful!") | log.Info("ORM engine initialization successful!") | ||||
| } else { | } else { | ||||
| log.Fatal("ORM engine initialization failed: %v", err) | log.Fatal("ORM engine initialization failed: %v", err) | ||||
| @@ -16,6 +16,7 @@ import ( | |||||
| "code.gitea.io/gitea/modules/base" | "code.gitea.io/gitea/modules/base" | ||||
| "code.gitea.io/gitea/modules/context" | "code.gitea.io/gitea/modules/context" | ||||
| "code.gitea.io/gitea/modules/generate" | "code.gitea.io/gitea/modules/generate" | ||||
| "code.gitea.io/gitea/modules/graceful" | |||||
| "code.gitea.io/gitea/modules/log" | "code.gitea.io/gitea/modules/log" | ||||
| "code.gitea.io/gitea/modules/setting" | "code.gitea.io/gitea/modules/setting" | ||||
| "code.gitea.io/gitea/modules/user" | "code.gitea.io/gitea/modules/user" | ||||
| @@ -351,7 +352,7 @@ func InstallPost(ctx *context.Context, form auth.InstallForm) { | |||||
| return | return | ||||
| } | } | ||||
| GlobalInit() | |||||
| GlobalInit(graceful.GetManager().HammerContext()) | |||||
| // Create admin account | // Create admin account | ||||
| if len(form.AdminName) > 0 { | if len(form.AdminName) > 0 { | ||||
| @@ -5,11 +5,14 @@ | |||||
| package mirror | package mirror | ||||
| import ( | import ( | ||||
| "context" | |||||
| "fmt" | "fmt" | ||||
| "net/url" | "net/url" | ||||
| "strings" | "strings" | ||||
| "time" | "time" | ||||
| "code.gitea.io/gitea/modules/graceful" | |||||
| "code.gitea.io/gitea/models" | "code.gitea.io/gitea/models" | ||||
| "code.gitea.io/gitea/modules/cache" | "code.gitea.io/gitea/modules/cache" | ||||
| "code.gitea.io/gitea/modules/git" | "code.gitea.io/gitea/modules/git" | ||||
| @@ -294,29 +297,38 @@ func Password(m *models.Mirror) string { | |||||
| } | } | ||||
| // Update checks and updates mirror repositories. | // Update checks and updates mirror repositories. | ||||
| func Update() { | |||||
| func Update(ctx context.Context) { | |||||
| log.Trace("Doing: Update") | log.Trace("Doing: Update") | ||||
| if err := models.MirrorsIterate(func(idx int, bean interface{}) error { | if err := models.MirrorsIterate(func(idx int, bean interface{}) error { | ||||
| m := bean.(*models.Mirror) | m := bean.(*models.Mirror) | ||||
| if m.Repo == nil { | if m.Repo == nil { | ||||
| log.Error("Disconnected mirror repository found: %d", m.ID) | log.Error("Disconnected mirror repository found: %d", m.ID) | ||||
| return nil | return nil | ||||
| } | } | ||||
| mirrorQueue.Add(m.RepoID) | |||||
| return nil | |||||
| select { | |||||
| case <-ctx.Done(): | |||||
| return fmt.Errorf("Aborted due to shutdown") | |||||
| default: | |||||
| mirrorQueue.Add(m.RepoID) | |||||
| return nil | |||||
| } | |||||
| }); err != nil { | }); err != nil { | ||||
| log.Error("Update: %v", err) | log.Error("Update: %v", err) | ||||
| } | } | ||||
| } | } | ||||
| // SyncMirrors checks and syncs mirrors. | // SyncMirrors checks and syncs mirrors. | ||||
| // TODO: sync more mirrors at same time. | |||||
| func SyncMirrors() { | |||||
| // FIXME: graceful: this should be a persistable queue | |||||
| func SyncMirrors(ctx context.Context) { | |||||
| // Start listening on new sync requests. | // Start listening on new sync requests. | ||||
| for repoID := range mirrorQueue.Queue() { | |||||
| syncMirror(repoID) | |||||
| for { | |||||
| select { | |||||
| case <-ctx.Done(): | |||||
| mirrorQueue.Close() | |||||
| return | |||||
| case repoID := <-mirrorQueue.Queue(): | |||||
| syncMirror(repoID) | |||||
| } | |||||
| } | } | ||||
| } | } | ||||
| @@ -416,7 +428,7 @@ func syncMirror(repoID string) { | |||||
| // InitSyncMirrors initializes a go routine to sync the mirrors | // InitSyncMirrors initializes a go routine to sync the mirrors | ||||
| func InitSyncMirrors() { | func InitSyncMirrors() { | ||||
| go SyncMirrors() | |||||
| go graceful.GetManager().RunWithShutdownContext(SyncMirrors) | |||||
| } | } | ||||
| // StartToMirror adds repoID to mirror queue | // StartToMirror adds repoID to mirror queue | ||||
| @@ -6,6 +6,7 @@ | |||||
| package pull | package pull | ||||
| import ( | import ( | ||||
| "context" | |||||
| "fmt" | "fmt" | ||||
| "io/ioutil" | "io/ioutil" | ||||
| "os" | "os" | ||||
| @@ -16,6 +17,7 @@ import ( | |||||
| "code.gitea.io/gitea/models" | "code.gitea.io/gitea/models" | ||||
| "code.gitea.io/gitea/modules/git" | "code.gitea.io/gitea/modules/git" | ||||
| "code.gitea.io/gitea/modules/graceful" | |||||
| "code.gitea.io/gitea/modules/log" | "code.gitea.io/gitea/modules/log" | ||||
| "code.gitea.io/gitea/modules/setting" | "code.gitea.io/gitea/modules/setting" | ||||
| "code.gitea.io/gitea/modules/sync" | "code.gitea.io/gitea/modules/sync" | ||||
| @@ -151,65 +153,53 @@ func manuallyMerged(pr *models.PullRequest) bool { | |||||
| // TestPullRequests checks and tests untested patches of pull requests. | // TestPullRequests checks and tests untested patches of pull requests. | ||||
| // TODO: test more pull requests at same time. | // TODO: test more pull requests at same time. | ||||
| func TestPullRequests() { | |||||
| prs, err := models.GetPullRequestsByCheckStatus(models.PullRequestStatusChecking) | |||||
| if err != nil { | |||||
| log.Error("Find Checking PRs: %v", err) | |||||
| return | |||||
| } | |||||
| var checkedPRs = make(map[int64]struct{}) | |||||
| func TestPullRequests(ctx context.Context) { | |||||
| // Update pull request status. | |||||
| for _, pr := range prs { | |||||
| checkedPRs[pr.ID] = struct{}{} | |||||
| if err := pr.GetBaseRepo(); err != nil { | |||||
| log.Error("GetBaseRepo: %v", err) | |||||
| continue | |||||
| } | |||||
| if manuallyMerged(pr) { | |||||
| continue | |||||
| go func() { | |||||
| prs, err := models.GetPullRequestIDsByCheckStatus(models.PullRequestStatusChecking) | |||||
| if err != nil { | |||||
| log.Error("Find Checking PRs: %v", err) | |||||
| return | |||||
| } | } | ||||
| if err := TestPatch(pr); err != nil { | |||||
| log.Error("testPatch: %v", err) | |||||
| continue | |||||
| for _, prID := range prs { | |||||
| select { | |||||
| case <-ctx.Done(): | |||||
| return | |||||
| default: | |||||
| pullRequestQueue.Add(prID) | |||||
| } | |||||
| } | } | ||||
| checkAndUpdateStatus(pr) | |||||
| } | |||||
| }() | |||||
| // Start listening on new test requests. | // Start listening on new test requests. | ||||
| for prID := range pullRequestQueue.Queue() { | |||||
| log.Trace("TestPullRequests[%v]: processing test task", prID) | |||||
| pullRequestQueue.Remove(prID) | |||||
| id := com.StrTo(prID).MustInt64() | |||||
| if _, ok := checkedPRs[id]; ok { | |||||
| continue | |||||
| } | |||||
| pr, err := models.GetPullRequestByID(id) | |||||
| if err != nil { | |||||
| log.Error("GetPullRequestByID[%s]: %v", prID, err) | |||||
| continue | |||||
| } else if manuallyMerged(pr) { | |||||
| continue | |||||
| } | |||||
| pr.Status = models.PullRequestStatusChecking | |||||
| if err := pr.Update(); err != nil { | |||||
| log.Error("testPatch[%d]: Unable to update status to Checking Status %v", pr.ID, err) | |||||
| continue | |||||
| } | |||||
| if err = TestPatch(pr); err != nil { | |||||
| log.Error("testPatch[%d]: %v", pr.ID, err) | |||||
| continue | |||||
| for { | |||||
| select { | |||||
| case prID := <-pullRequestQueue.Queue(): | |||||
| log.Trace("TestPullRequests[%v]: processing test task", prID) | |||||
| pullRequestQueue.Remove(prID) | |||||
| id := com.StrTo(prID).MustInt64() | |||||
| pr, err := models.GetPullRequestByID(id) | |||||
| if err != nil { | |||||
| log.Error("GetPullRequestByID[%s]: %v", prID, err) | |||||
| continue | |||||
| } else if manuallyMerged(pr) { | |||||
| continue | |||||
| } else if err = TestPatch(pr); err != nil { | |||||
| log.Error("testPatch[%d]: %v", pr.ID, err) | |||||
| continue | |||||
| } | |||||
| checkAndUpdateStatus(pr) | |||||
| case <-ctx.Done(): | |||||
| pullRequestQueue.Close() | |||||
| log.Info("PID: %d Pull Request testing shutdown", os.Getpid()) | |||||
| return | |||||
| } | } | ||||
| checkAndUpdateStatus(pr) | |||||
| } | } | ||||
| } | } | ||||
| // Init runs the task queue to test all the checking status pull requests | // Init runs the task queue to test all the checking status pull requests | ||||
| func Init() { | func Init() { | ||||
| go TestPullRequests() | |||||
| go graceful.GetManager().RunWithShutdownContext(TestPullRequests) | |||||
| } | } | ||||
| @@ -5,12 +5,14 @@ | |||||
| package pull | package pull | ||||
| import ( | import ( | ||||
| "context" | |||||
| "fmt" | "fmt" | ||||
| "os" | "os" | ||||
| "path" | "path" | ||||
| "code.gitea.io/gitea/models" | "code.gitea.io/gitea/models" | ||||
| "code.gitea.io/gitea/modules/git" | "code.gitea.io/gitea/modules/git" | ||||
| "code.gitea.io/gitea/modules/graceful" | |||||
| "code.gitea.io/gitea/modules/log" | "code.gitea.io/gitea/modules/log" | ||||
| "code.gitea.io/gitea/modules/notification" | "code.gitea.io/gitea/modules/notification" | ||||
| issue_service "code.gitea.io/gitea/services/issue" | issue_service "code.gitea.io/gitea/services/issue" | ||||
| @@ -54,6 +56,7 @@ func checkForInvalidation(requests models.PullRequestList, repoID int64, doer *m | |||||
| return fmt.Errorf("git.OpenRepository: %v", err) | return fmt.Errorf("git.OpenRepository: %v", err) | ||||
| } | } | ||||
| go func() { | go func() { | ||||
| // FIXME: graceful: We need to tell the manager we're doing something... | |||||
| err := requests.InvalidateCodeComments(doer, gitRepo, branch) | err := requests.InvalidateCodeComments(doer, gitRepo, branch) | ||||
| if err != nil { | if err != nil { | ||||
| log.Error("PullRequestList.InvalidateCodeComments: %v", err) | log.Error("PullRequestList.InvalidateCodeComments: %v", err) | ||||
| @@ -79,39 +82,45 @@ func addHeadRepoTasks(prs []*models.PullRequest) { | |||||
| // and generate new patch for testing as needed. | // and generate new patch for testing as needed. | ||||
| func AddTestPullRequestTask(doer *models.User, repoID int64, branch string, isSync bool) { | func AddTestPullRequestTask(doer *models.User, repoID int64, branch string, isSync bool) { | ||||
| log.Trace("AddTestPullRequestTask [head_repo_id: %d, head_branch: %s]: finding pull requests", repoID, branch) | log.Trace("AddTestPullRequestTask [head_repo_id: %d, head_branch: %s]: finding pull requests", repoID, branch) | ||||
| prs, err := models.GetUnmergedPullRequestsByHeadInfo(repoID, branch) | |||||
| if err != nil { | |||||
| log.Error("Find pull requests [head_repo_id: %d, head_branch: %s]: %v", repoID, branch, err) | |||||
| return | |||||
| } | |||||
| graceful.GetManager().RunWithShutdownContext(func(ctx context.Context) { | |||||
| // There is no sensible way to shut this down ":-(" | |||||
| // If you don't let it run all the way then you will lose data | |||||
| // FIXME: graceful: AddTestPullRequestTask needs to become a queue! | |||||
| if isSync { | |||||
| requests := models.PullRequestList(prs) | |||||
| if err = requests.LoadAttributes(); err != nil { | |||||
| log.Error("PullRequestList.LoadAttributes: %v", err) | |||||
| } | |||||
| if invalidationErr := checkForInvalidation(requests, repoID, doer, branch); invalidationErr != nil { | |||||
| log.Error("checkForInvalidation: %v", invalidationErr) | |||||
| prs, err := models.GetUnmergedPullRequestsByHeadInfo(repoID, branch) | |||||
| if err != nil { | |||||
| log.Error("Find pull requests [head_repo_id: %d, head_branch: %s]: %v", repoID, branch, err) | |||||
| return | |||||
| } | } | ||||
| if err == nil { | |||||
| for _, pr := range prs { | |||||
| pr.Issue.PullRequest = pr | |||||
| notification.NotifyPullRequestSynchronized(doer, pr) | |||||
| if isSync { | |||||
| requests := models.PullRequestList(prs) | |||||
| if err = requests.LoadAttributes(); err != nil { | |||||
| log.Error("PullRequestList.LoadAttributes: %v", err) | |||||
| } | |||||
| if invalidationErr := checkForInvalidation(requests, repoID, doer, branch); invalidationErr != nil { | |||||
| log.Error("checkForInvalidation: %v", invalidationErr) | |||||
| } | |||||
| if err == nil { | |||||
| for _, pr := range prs { | |||||
| pr.Issue.PullRequest = pr | |||||
| notification.NotifyPullRequestSynchronized(doer, pr) | |||||
| } | |||||
| } | } | ||||
| } | } | ||||
| } | |||||
| addHeadRepoTasks(prs) | |||||
| addHeadRepoTasks(prs) | |||||
| log.Trace("AddTestPullRequestTask [base_repo_id: %d, base_branch: %s]: finding pull requests", repoID, branch) | |||||
| prs, err = models.GetUnmergedPullRequestsByBaseInfo(repoID, branch) | |||||
| if err != nil { | |||||
| log.Error("Find pull requests [base_repo_id: %d, base_branch: %s]: %v", repoID, branch, err) | |||||
| return | |||||
| } | |||||
| for _, pr := range prs { | |||||
| AddToTaskQueue(pr) | |||||
| } | |||||
| log.Trace("AddTestPullRequestTask [base_repo_id: %d, base_branch: %s]: finding pull requests", repoID, branch) | |||||
| prs, err = models.GetUnmergedPullRequestsByBaseInfo(repoID, branch) | |||||
| if err != nil { | |||||
| log.Error("Find pull requests [base_repo_id: %d, base_branch: %s]: %v", repoID, branch, err) | |||||
| return | |||||
| } | |||||
| for _, pr := range prs { | |||||
| AddToTaskQueue(pr) | |||||
| } | |||||
| }) | |||||
| } | } | ||||
| // PushToBaseRepo pushes commits from branches of head repository to | // PushToBaseRepo pushes commits from branches of head repository to | ||||