2 ###############################################################################
7 mig - git-like program for tracking and manipulating legacy data files for
12 B<mig> <command> [argument] [...]
16 B<mig> is used to track and manipulate CSV or CSV-like text files exported from
17 legacy systems for migration into Evergreen. It can be a wrapper for some
18 other migration tools and tracks state using a PostgreSQL table in a given
21 It makes use of certain environment variables that may be set by the B<mig-env>
22 tool: PGHOST, PGPORT, PGUSER, PGDATABASE, MIGSCHEMA, and MIGWORKDIR
24 For most commands, if the current working directory falls outside of the
25 directory specified by MIGWORKDIR, then mig will assume that environment is
26 also incorrect and bail before doing any actual work.
28 ~/.pgpass should also be configured, as B<mig> will not prompt for a database
31 Only the B<env> and B<help> commands work without the MIGSCHEMA environment
36 Using B<mig> should go something like this:
40 =item mig env create m_foo # Sets up the environment
42 =item mig env use m_foo # Spawns a shell using the configured environment
44 =item mig init # creates the m_foo schema in the database if needed, and other tables
46 =item mig add patrons.tsv # tracks an incoming data file; repeat for additional files
48 =item mig iconv patrons.tsv # convert it to UTF8, creating patrons.tsv.utf8
50 =item mig clean patrons.tsv # cleans the file, creating patrons.tsv.utf8.clean
52 =item mig link patrons.tsv actor_usr # makes the soon-to-be staging table a child of m_foo.actor_usr
54 =item mig convert patrons.tsv # creates a .sql file for staging the data
56 =item mig stage patrons.tsv # load said .sql file
58 =item mig mapper patrons.tsv # interactive tool for analyzing/mapping the staging table
60 =item mig analysis patrons.tsv # writes a summary .tsv file of mapped/flagged fields from the staging table
62 =item mig map patrons.tsv # apply configured mappings
64 =item mig write_prod patrons.tsv # creates a .sql file for pushing the staging data into production
66 =item mig reporter --analyst "Foo Fooer" --report_title "Foo Load Analysis" #creates an asciidoc report
68 =item mig gsheet --pull foo_tab_name OR --push foo_pg_table_name
76 =item B<help> [command]
78 Display this very same documentation, or specific documentation for one of the
81 =item B<env> <create|use|show> <schema>
83 Invokes B<mig-env> with the same arguments. I<mig-env> can set important
84 environment variables and spawn a shell with those variables, and it also does
85 some directory creation and symlinking.
89 Create or re-create the PostgreSQL tracking table for the schema specified by
90 the MIGSCHEMA environment variable. If needed, create the migration schema
91 itself and run migration_tools.init() and build() if the migration_tools schema
94 =item B<status> [file] [...]
96 Show status information for either the specified files or all tracked files if
99 =item B<add> [--no-headers|--headers] <file> [file|--no-headers|--headers] [...]
101 Add the specified files to the migration tracker. Until --no-headers is
102 specified, the tracker will assume the files have headers.
104 You can do crazy stuff like
105 B<mig add file1 --no-headers file2 file3 --headers file4>
107 =item B<remove> <file> [file] [...]
109 Remove the specified files from the migration tracker.
111 =item B<iconv> <file> [other arguments...]
113 Attempts to invoke B<iconv> on the specified tracked file, placing the output in
116 If given no other arguments, the invocation will lool like
120 iconv -f ISO-8859-1 -t UTF-8 -o <file>.utf8 <file>
124 otherwise, the arguments will be passed through like so
128 iconv [other arguments...] -o <file>.utf8 <file>
132 =item B<skip-iconv> <file>
134 If this is used instead of B<iconv>, then B<mig> will look for an existing
135 <file>.utf8 and use it instead of attempting to create one.
137 =item B<clean> <file> [other arguments...]
139 Attempts to invoke B<clean_csv> on the iconv-converted specified tracked file,
140 placing the output in <file>.utf8.clean
142 If given no other arguments, the invocation will lool like
146 clean_csv --config scripts/clean.conf --fix --apply <--create-headers> <file>
150 otherwise, the arguments will be passed through like so
154 clean_csv [other arguments...] <file>
158 =item B<skip-clean> <file>
160 If this is used instead of B<clean>, then B<mig> will look for an existing
161 <file>.utf8.clean and use it instead of attempting to create one.
163 =item B<link> <file> <parent table>
165 Associate the specified file with a parent table within the migration schema.
167 Linking multiple files to the same parent table is not allowed currently.
169 =item B<unlink> <file>
171 Removes any association between the specified file and a parent table within
172 the migration schema.
174 =item B<convert> <file>
176 Attempts to invoke B<csv2sql> on the .utf8.clean version of the specified
177 tracked file, creating either [file].utf8.clean.stage.sql or
178 <parent table>_stage.sql depending on whether the file has been linked to a
179 parent table within the migration schema or not.
181 If given no other arguments, the invocation will lool like
185 csv2sql --config scripts/clean.conf --add-x-migrate --schema <MIGSCHEMA> [--parent <PARENT TABLE>] -o <[<file>.utf8.clean.stage.sql]|[parent_table_stage.sql]> <FILE>.utf8.clean
189 otherwise, the arguments will be passed through like so
193 csv2sql [other arguments...] -o <[<file>.utf8.clean.stage.sql]|[parent_table_stage.sql]> <file>.utf8.clean
197 =item B<stage> <file> [other arguments...]
199 Load the SQL-converted version of the specified file into the migration schema.
201 Extra arguments are passed to the underlying call to psql
203 =item B<mapper> <file>
205 Interactive session for analyzing, flagging, and mapping legacy field data to
208 Upon exit, generate either [file].clean.map.sql or <parent table>_map.sql. The
209 SQL generated will be UPDATE's for setting the Evergreen-specific columns for a
210 given file's staging tables, and TRUNCATE's and INSERT's for auxilary tables.
211 The files will have \include hooks for pulling in additional mapping files
212 (for example, end-user mappings for circ modifiers, etc.)
214 =item B<analysis> [file]
216 Writes a MIGSCHEMA.tsv file containing a break-down of mapped and flagged
217 fields from the specified file, or all staged files if no file is specified.
219 The main goal of the tsv file is to present end-user mappable data for circ
220 modifiers, shelving locations, patron profiles, etc. We use spreadsheets for
221 this now but may move to a dedicated UI in the future.
225 Applies the mapping sql to the migration schema for the specified mapped file,
226 or for all mapped files if no file is specified.
228 =item B<write_prod> [file]
230 Generates <parent table>_prod.sql for the specified linked and mapped file, or
231 all such files if no file is specified.
233 =item B<sql> [arguments...]
235 A wrapper around the psql command. At some point the plan is to shove mig-tracked variables into psql sessions.
237 =item B<reporter> --analyst "Analyst Name" --report_title "Report Title"
239 Generates an asciidoc file in the git working directory that can be converted to
240 any appropriate format. The analyst and report parameters are required.
242 Optional parameters are :
244 --added_page_title and --added_page_file
246 If one is used both must be. The added page file can be plain text or asciidoc. This
247 adds an extra arbitrary page of notes to the report. Mig assumes the page file is in the mig git directory.
251 This will define a set of tags to use, if not set it will default to Circs,
252 Holds, Actors, Bibs, Assets & Money.
256 Allows you to override the default evergreen_staged_report.xml in the mig-xml folder.
258 =item B<gsheet> --pull spreadsheet_tab or --push postgres_table
260 This uses the gsheet_tracked_table and gsheet_tracked column tables to map a Google Docs Spreadsheet tabs
261 with Postgres tables in the mig schema. Multiple spreadsheets can be used but tab names must be unique.
262 Each spreadsheet column needs a header that matches the column name in the matching table. An oauth
263 session key is also needed for your Google account and mig gsheet will look for it in the .mig directory.
269 ###############################################################################
274 HOME PGHOST PGPORT PGUSER PGDATABASE MIGSCHEMA
275 MIGBASEWORKDIR MIGBASEGITDIR MIGGITDIR MIGWORKDIR
279 my $mig_bin = "$FindBin::Bin/mig-bin/";
280 use lib "$FindBin::Bin/mig-bin";
283 pod2usage(-verbose => 2) if ! $ARGV[0];
286 if (defined $ARGV[1]) {
287 my $cmd = $mig_bin . "mig-$ARGV[1]";
289 system( $mig_bin . "mig-$ARGV[1]", '--help' );
291 pod2usage(-verbose => 2);
294 pod2usage(-verbose => 2);
298 standard_invocation(@ARGV);
301 Mig::die_if_no_env_migschema();
302 standard_invocation(@ARGV);
305 Mig::die_if_no_env_migschema();
306 standard_invocation(@ARGV);
309 Mig::die_if_no_env_migschema();
310 standard_invocation(@ARGV);
313 Mig::die_if_no_env_migschema();
314 standard_invocation(@ARGV);
317 Mig::die_if_no_env_migschema();
318 standard_invocation(@ARGV);
321 Mig::die_if_no_env_migschema();
322 standard_invocation(@ARGV);
325 Mig::die_if_no_env_migschema();
326 standard_invocation(@ARGV);
329 Mig::die_if_no_env_migschema();
330 standard_invocation(@ARGV);
333 Mig::die_if_no_env_migschema();
334 standard_invocation(@ARGV);
337 Mig::die_if_no_env_migschema();
338 standard_invocation(@ARGV);
341 Mig::die_if_no_env_migschema();
342 standard_invocation(@ARGV);
345 Mig::die_if_no_env_migschema();
346 standard_invocation(@ARGV);
349 Mig::die_if_no_env_migschema();
350 standard_invocation(@ARGV);
353 Mig::die_if_no_env_migschema();
354 standard_invocation(@ARGV);
357 Mig::die_if_no_env_migschema();
358 standard_invocation(@ARGV);
361 Mig::die_if_no_env_migschema();
362 standard_invocation(@ARGV);
365 Mig::die_if_no_env_migschema();
366 standard_invocation(@ARGV);
369 Mig::die_if_no_env_migschema();
372 Mig::die_if_no_env_migschema();
375 print "$MIGWORKDIR\n";
378 print "$MIGBASEGITDIR\n";
381 print "$MIGGITDIR\n";
388 sub standard_invocation {
390 system( $mig_bin . "mig-$cmd", @_ );