Uploaded image for project: 'Crowd Data Center'
  1. Crowd Data Center
  2. CWD-1212

Improving memory usage when importing large datasets

    XMLWordPrintable

Details

    • Our product teams collect and evaluate feedback from a number of different sources. To learn more about how we use customer feedback in the planning process, check out our new feature policy.

    Description

      Currently, importing large sets of data requires a lot of memory as the collection of data being imported is loaded up to memory before being purged to the database. In order to improve memory efficiency, we need to perform the import in "chunks" such that we read a chunk from the input and write a chunk out to the database. The writing process already supports purging in "chunks" via JDBC batch processing. The input process will need to read in chunks in various ways:

      • XML import: use event-based XML processing
      • CSV import: read data as a stream
      • JDBC import: read pages of data

      Attachments

        Issue Links

          Activity

            People

              Unassigned Unassigned
              shamid@atlassian.com shihab
              Votes:
              2 Vote for this issue
              Watchers:
              5 Start watching this issue

              Dates

                Created:
                Updated: