Skip to content

How to bring in Candidate URLs from the test server

Ashish Acharya edited this page May 31, 2023 · 9 revisions

Bringing in select collections

  1. Log into the test server (bastion)

  2. From the bastion, RDP into the following server 10.51.14.135

  3. On the server, open up Anaconda Powershell Prompt (miniconda3) and navigate to E:\sde-indexing-helper\config_generation

  4. Find python file called export_collections.py and, in the list called COLLECTIONS_TO_UPLOAD, add a list of collection config folders.

  5. Run python export_collections.py

  6. This will upload zipped json files to the S3 bucket here.

  7. Go to the collections admin page here.

  8. Select collection ONE OR TWO AT A TIME (otherwise celery might throw an error because each process tries to write to postgres at the same time). Then from the dropdown menu click on Import candidate URLs

Bringing in everything on the test server

Follow steps, 1, 2, 3 from above.

  1. Run python export_whole_index.py

  2. This will upload one large zipped json file to the S3 bucket here.

  3. Log into the sde_indexing_helper server and navigate to the folder sde_indexing_helper. Run drun /bin/bash. Then inside the docker run python manage.py load_urls_from_s3

If you spot something in the pipeline or this documentation that can be optimized, be the change you wish to see in the world. :D