We have a private sftp file server where we have access to a number of the bulk data products. The ones we are using is the 195/198, 208/209, 230/233 files.
We also have access to the accompanying schema info.
We ingest the snapshot files into our DB and are building an event driven architecture that ensure that the DB stays up to date, ie it gets the update file from the CH SFTP server, and updates the db accordingly.
For the 195/198 and the 230/233 product, this is working.
However, we have a few different questions.
- Does the update type matter for 209? What do the fields mean, and do they match the update types in 198?
- What is the recommended way to update this, if we are loop through the dates to update snapshot, can we apply all the 198 updates from snapshot date → now and THEN apply all the 209 updates from snapshot date → now or do we need to apply them sequentially 198 → 209 → 198 - 209 from snapshot date → now?
- We have seen on the 195/198 data, all external numbers will be fixed to 12 digits. However for the 208/209 data, in the data sheet, it says leading zeros will be removed. We just wanted to understand why? Should we add them back to then connect the datasets?
- Given the 195/198 and 208/209 are implicitly linked we need to get snapshot files from the exact same day to ensure that we can keep these datasets up to date, however on the file server we have access to, only the 195 snapshots drop regularly, our 208 snapshot is more than 3 months old.
We were interacting with some other team at CH but they claim they can no longer help us. This team included Paula Williams, Patrick Hughes, William Noble, Kevin Watts and Venkaiah Nara among others.
Please can you point me to the correct team/group of people that are able to help us.
FYI we are a contractor working for a government client department.
Thanks,
Daniel