| Actions items Action items (Recap & background Where to source data from? Revisiting action items from 7/15/2024 ):- Populate spreadsheet
Spreadsheet of reference data (AF) - Develop a
Google Form for SIG responses (AF). This will tell us which types of reference data we need to create ourselves and which types we can sourceto collect both sample record data and sample reference data types Check about effort needed to get a fresh export from Chicago and documentation on the process (TO) Set up next meeting (YK)
Identify types of data used in each of the apps as well as data at the tenant level - move the table to spreadsheet Reference data by app Loan types, fund codes, call number types, etc.
Record types by app Orders, order lines, users, instance/holdings/items, etc.
App settings Data Import job profiles, Inventory export targets, etc.
Tenant-wide data & settings Libraries, location codes, service points Consortial partners & relationships Permission sets
Insert links to GitHub repositories – CW - is this still relevant Also solicit data samples from respective libraries - e.g. Order data as mentioned by Maccabee Levine (Lehigh) Bound-with (one item linking to multiple holdings (GBV)
| | Documents: How did Chicago share data? Exported out. Inventory data is relatively safe for sharing How did we officially share the data? Can we share bigger sets of data for BugFest, and then create a smaller set for Snapshot? Chicago data was originally scoped to Inventory, and did not contain connections to other records. How do we deal with different institutions' data? What does the environment that we are putting the data into need to look like? What should the FOLIO BugFest actually be managed Maybe we need to carefully curate the reference data in advance of getting the data Libraries should be testing the logical portion of features, and not full workflows We’ll want to iteratively update BugFest dataset across different areas What’s in BugFest right now? Verify different types of data and what is working for people?
Do we need a consortial data set? | 5 min
| Future meeting times
| Find a regular meeting time - once a week, or bi-weeklyRecap of where we are: Plan is to set up a blank environment Ask SMEs and users to upload sample data that they need for testing Take a snapshot and use as a golden copy Will need to ensure ongoing maintenance of this environment as features and apps are built out and require new sample data
Where to source sample data Chicago’s data set uses a customized MARC mapping rather than the default; Chicago is also not using MARC authority data We need a library using ERM, MARC authorities, and default MARC mapping Robust anonymization will be required. Lee’s plan: Replace PII with randomly generated data Scramble loan history Scramble orders, invoice amounts, fund codes Replace vendor names with randomized names Strip out staff notes with initials, etc.
One set of data for the general environment, and perhaps a second sample set for the ECS environment
7/15 action items Spreadsheet has been populated with all modules and their related SIGs This is what we will use to compile and deliver the final dataset to devs
Form has been drafted to solicit input from SIGs, SMEs, POs, etc. Yogesh and Lee will review and let Autumn know about correx
Getting sample data sets Autumn and Tod will bring write-up of Lee’s proposal to administrations, check on feasibility and willingness to use Chicago and MSU data sets
Meetings will be every Tuesday at 6pm CET
For next time: |