backup.go 8.5 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323
  1. package database
  2. import (
  3. "bufio"
  4. "bytes"
  5. "context"
  6. "fmt"
  7. "io"
  8. "os"
  9. "path/filepath"
  10. "reflect"
  11. "strings"
  12. "sync"
  13. "github.com/cockroachdb/errors"
  14. jsoniter "github.com/json-iterator/go"
  15. "gorm.io/gorm"
  16. "gorm.io/gorm/schema"
  17. log "unknwon.dev/clog/v2"
  18. "gogs.io/gogs/internal/conf"
  19. "gogs.io/gogs/internal/osutil"
  20. )
  21. // getTableType returns the type name of a table definition without package name,
  22. // e.g. *database.LFSObject -> LFSObject.
  23. func getTableType(t any) string {
  24. return strings.TrimPrefix(fmt.Sprintf("%T", t), "*database.")
  25. }
  26. // DumpDatabase dumps all data from database to file system in JSON Lines format.
  27. func DumpDatabase(ctx context.Context, db *gorm.DB, dirPath string, verbose bool) error {
  28. err := os.MkdirAll(dirPath, os.ModePerm)
  29. if err != nil {
  30. return err
  31. }
  32. err = dumpLegacyTables(ctx, dirPath, verbose)
  33. if err != nil {
  34. return errors.Wrap(err, "dump legacy tables")
  35. }
  36. for _, table := range Tables {
  37. select {
  38. case <-ctx.Done():
  39. return ctx.Err()
  40. default:
  41. }
  42. tableName := getTableType(table)
  43. if verbose {
  44. log.Trace("Dumping table %q...", tableName)
  45. }
  46. err := func() error {
  47. tableFile := filepath.Join(dirPath, tableName+".json")
  48. f, err := os.Create(tableFile)
  49. if err != nil {
  50. return errors.Wrap(err, "create table file")
  51. }
  52. defer func() { _ = f.Close() }()
  53. return dumpTable(ctx, db, table, f)
  54. }()
  55. if err != nil {
  56. return errors.Wrapf(err, "dump table %q", tableName)
  57. }
  58. }
  59. return nil
  60. }
  61. func dumpTable(ctx context.Context, db *gorm.DB, table any, w io.Writer) error {
  62. query := db.WithContext(ctx).Model(table)
  63. switch table.(type) {
  64. case *LFSObject:
  65. query = query.Order("repo_id, oid ASC")
  66. default:
  67. query = query.Order("id ASC")
  68. }
  69. rows, err := query.Rows()
  70. if err != nil {
  71. return errors.Wrap(err, "select rows")
  72. }
  73. defer func() { _ = rows.Close() }()
  74. for rows.Next() {
  75. elem := reflect.New(reflect.TypeOf(table).Elem()).Interface()
  76. err = db.ScanRows(rows, elem)
  77. if err != nil {
  78. return errors.Wrap(err, "scan rows")
  79. }
  80. switch e := elem.(type) {
  81. case *LFSObject:
  82. e.CreatedAt = e.CreatedAt.UTC()
  83. }
  84. err = jsoniter.NewEncoder(w).Encode(elem)
  85. if err != nil {
  86. return errors.Wrap(err, "encode JSON")
  87. }
  88. }
  89. return rows.Err()
  90. }
  91. func dumpLegacyTables(ctx context.Context, dirPath string, verbose bool) error {
  92. // Purposely create a local variable to not modify global variable
  93. legacyTables := append(legacyTables, new(Version))
  94. for _, table := range legacyTables {
  95. select {
  96. case <-ctx.Done():
  97. return ctx.Err()
  98. default:
  99. }
  100. tableName := getTableType(table)
  101. if verbose {
  102. log.Trace("Dumping table %q...", tableName)
  103. }
  104. err := func() error {
  105. tableFile := filepath.Join(dirPath, tableName+".json")
  106. f, err := os.Create(tableFile)
  107. if err != nil {
  108. return errors.Wrap(err, "create JSON file")
  109. }
  110. defer func() { _ = f.Close() }()
  111. return dumpTable(ctx, db, table, f)
  112. }()
  113. if err != nil {
  114. return errors.Wrapf(err, "dump table %q", tableName)
  115. }
  116. }
  117. return nil
  118. }
  119. // ImportDatabase imports data from backup archive in JSON Lines format.
  120. func ImportDatabase(ctx context.Context, db *gorm.DB, dirPath string, verbose bool) error {
  121. err := importLegacyTables(ctx, dirPath, verbose)
  122. if err != nil {
  123. return errors.Wrap(err, "import legacy tables")
  124. }
  125. for _, table := range Tables {
  126. select {
  127. case <-ctx.Done():
  128. return ctx.Err()
  129. default:
  130. }
  131. tableName := strings.TrimPrefix(fmt.Sprintf("%T", table), "*database.")
  132. err := func() error {
  133. tableFile := filepath.Join(dirPath, tableName+".json")
  134. if !osutil.IsFile(tableFile) {
  135. log.Info("Skipped table %q", tableName)
  136. return nil
  137. }
  138. if verbose {
  139. log.Trace("Importing table %q...", tableName)
  140. }
  141. f, err := os.Open(tableFile)
  142. if err != nil {
  143. return errors.Wrap(err, "open table file")
  144. }
  145. defer func() { _ = f.Close() }()
  146. return importTable(ctx, db, table, f)
  147. }()
  148. if err != nil {
  149. return errors.Wrapf(err, "import table %q", tableName)
  150. }
  151. }
  152. return nil
  153. }
  154. func importTable(ctx context.Context, db *gorm.DB, table any, r io.Reader) error {
  155. err := db.WithContext(ctx).Migrator().DropTable(table)
  156. if err != nil {
  157. return errors.Wrap(err, "drop table")
  158. }
  159. err = db.WithContext(ctx).Migrator().AutoMigrate(table)
  160. if err != nil {
  161. return errors.Wrap(err, "auto migrate")
  162. }
  163. s, err := schema.Parse(table, &sync.Map{}, db.NamingStrategy)
  164. if err != nil {
  165. return errors.Wrap(err, "parse schema")
  166. }
  167. rawTableName := s.Table
  168. skipResetIDSeq := map[string]bool{
  169. "lfs_object": true,
  170. }
  171. scanner := bufio.NewScanner(r)
  172. for scanner.Scan() {
  173. // PostgreSQL does not like the null characters (U+0000)
  174. cleaned := bytes.ReplaceAll(scanner.Bytes(), []byte("\\u0000"), []byte(""))
  175. elem := reflect.New(reflect.TypeOf(table).Elem()).Interface()
  176. err = jsoniter.Unmarshal(cleaned, elem)
  177. if err != nil {
  178. return errors.Wrap(err, "unmarshal JSON to struct")
  179. }
  180. err = db.WithContext(ctx).Create(elem).Error
  181. if err != nil {
  182. return errors.Wrap(err, "create row")
  183. }
  184. }
  185. // PostgreSQL needs manually reset table sequence for auto increment keys
  186. if conf.UsePostgreSQL && !skipResetIDSeq[rawTableName] {
  187. seqName := rawTableName + "_id_seq"
  188. if err = db.WithContext(ctx).Exec(fmt.Sprintf(`SELECT setval('%s', COALESCE((SELECT MAX(id)+1 FROM "%s"), 1), false)`, seqName, rawTableName)).Error; err != nil {
  189. return errors.Wrapf(err, "reset table %q.%q", rawTableName, seqName)
  190. }
  191. }
  192. return nil
  193. }
  194. func importLegacyTables(ctx context.Context, dirPath string, verbose bool) error {
  195. skipInsertProcessors := map[string]bool{
  196. "mirror": true,
  197. "milestone": true,
  198. }
  199. // Purposely create a local variable to not modify global variable
  200. legacyTables := append(legacyTables, new(Version))
  201. for _, table := range legacyTables {
  202. select {
  203. case <-ctx.Done():
  204. return ctx.Err()
  205. default:
  206. }
  207. tableName := strings.TrimPrefix(fmt.Sprintf("%T", table), "*database.")
  208. tableFile := filepath.Join(dirPath, tableName+".json")
  209. if !osutil.IsFile(tableFile) {
  210. continue
  211. }
  212. if verbose {
  213. log.Trace("Importing table %q...", tableName)
  214. }
  215. if err := db.WithContext(ctx).Migrator().DropTable(table); err != nil {
  216. return errors.Newf("drop table %q: %v", tableName, err)
  217. } else if err = db.WithContext(ctx).Migrator().AutoMigrate(table); err != nil {
  218. return errors.Newf("sync table %q: %v", tableName, err)
  219. }
  220. f, err := os.Open(tableFile)
  221. if err != nil {
  222. return errors.Newf("open JSON file: %v", err)
  223. }
  224. s, err := schema.Parse(table, &sync.Map{}, db.NamingStrategy)
  225. if err != nil {
  226. _ = f.Close()
  227. return errors.Wrap(err, "parse schema")
  228. }
  229. rawTableName := s.Table
  230. _, isInsertProcessor := table.(interface{ BeforeCreate(*gorm.DB) error })
  231. scanner := bufio.NewScanner(f)
  232. for scanner.Scan() {
  233. // PostgreSQL does not like the null characters (U+0000)
  234. cleaned := bytes.ReplaceAll(scanner.Bytes(), []byte("\\u0000"), []byte(""))
  235. if err = jsoniter.Unmarshal(cleaned, table); err != nil {
  236. _ = f.Close()
  237. return errors.Newf("unmarshal to struct: %v", err)
  238. }
  239. if err = db.WithContext(ctx).Create(table).Error; err != nil {
  240. _ = f.Close()
  241. return errors.Newf("insert struct: %v", err)
  242. }
  243. var meta struct {
  244. ID int64
  245. CreatedUnix int64
  246. DeadlineUnix int64
  247. ClosedDateUnix int64
  248. }
  249. if err = jsoniter.Unmarshal(cleaned, &meta); err != nil {
  250. log.Error("Failed to unmarshal to map: %v", err)
  251. }
  252. // Reset created_unix back to the date saved in archive because Create method updates its value
  253. if isInsertProcessor && !skipInsertProcessors[rawTableName] {
  254. if err = db.WithContext(ctx).Exec("UPDATE `"+rawTableName+"` SET created_unix=? WHERE id=?", meta.CreatedUnix, meta.ID).Error; err != nil {
  255. log.Error("Failed to reset '%s.created_unix': %v", rawTableName, err)
  256. }
  257. }
  258. switch rawTableName {
  259. case "milestone":
  260. if err = db.WithContext(ctx).Exec("UPDATE `"+rawTableName+"` SET deadline_unix=?, closed_date_unix=? WHERE id=?", meta.DeadlineUnix, meta.ClosedDateUnix, meta.ID).Error; err != nil {
  261. log.Error("Failed to reset 'milestone.deadline_unix', 'milestone.closed_date_unix': %v", err)
  262. }
  263. }
  264. }
  265. _ = f.Close()
  266. // PostgreSQL needs manually reset table sequence for auto increment keys
  267. if conf.UsePostgreSQL {
  268. seqName := rawTableName + "_id_seq"
  269. if err = db.WithContext(ctx).Exec(fmt.Sprintf(`SELECT setval('%s', COALESCE((SELECT MAX(id)+1 FROM "%s"), 1), false);`, seqName, rawTableName)).Error; err != nil {
  270. return errors.Newf("reset table %q' sequence: %v", rawTableName, err)
  271. }
  272. }
  273. }
  274. return nil
  275. }