While I still needed information about the missing columns in the CSV files found in the NutritionIX DB, I went ahead and added the columns I had access to. These included brand_name, ss_metric_qty, ss_metric_unit, item_id. However, the insertion process took a lot longer than it should have. I have some theories:
- The insertion is being done in a Python for-loop. This might be causing a lot of overhead.
- I am adding one row at a time instead of all at once
- Indexing in PostgreSQL
Please let me know what you think. I will be doing my own research with another python script.
Top comments (0)