The annual CUCC expedition to Austria lasts around five weeks each summer
and concentrates mainly on new exploration in the high alpine karst of the
Loser plateau (about 80km E of Salzburg). Recent large projects are:
Data Storage
- Version-controlled repositories
- website ('expoweb')
- troggle
- loser
- drawings (tunneldata)
- expofiles - just files
- photos
- presentations
- posters
- maps
- documents
Website Structure [as in 2015]
- Static web pages
- indexes, festering, background, handbook<
- Generated web pages
- Troggle pages
- cave pages,survey scans, logbook entries, triplists
- Troggle also 'wraps' the other [static] stuff
WEBSITE LAYOUT [as in 2015]
-
years
has per-year info (reports, lists, policy, logbooks)
-
handbook
contains handbook
- 1626, 1623 are the [Austrian kataster] areas
-
noinfo
- dir used to be hidden on public web
- scripts are in noinfo/ [or scripts/ ]
- generated files are regenerated with 'Make' in expoweb dir
(click on 'others'-> 'Regenerate website' is the same)
- prospecting guide is in noinfo/prospecting_guide_scripts [broken]
- folk is people list, now generated from folk/folk.csv by scripts/make-folklist.py
Editing [as in 2015]
Don't try to edit generated files.
Can edit in various ways:
- In expo file manager (best)
- Live on website (not recommended)
- (Clone and) edit on another computer
You always have to commit [version control] afterwards, however you edited.
Managing Files
EVERYTHING should be on the website!
(HUGE BINARIES SHOULD NOT BE IN version control)
Info should go in the website handbook (as html, or PDF)
Layout files edited every year make sense as ODT docs:
(survey index, labels, accounting procedure)
Link from website, store in website or expofiles/uploads
Don't just create a random doc and leave it in 'Desktop'.
Think about where some info would be found
Survex Dataset
caves
surface
surface/terrain
fixedpts
template: example files
docs/datamaintenance.txt has some rules.
Should be in the handbook
surface/caves connected by ents.svx, which is generated by make-ents.py, which
fishes out every point marked with *entrance in the dataset. ARGE do this wrong, with *fix in cave
all* make various groupings of data
Adding a CAVE
- Location
- Photo
- Survey data
- Survey notes
- Description
- Tag status
- Question marks
- Drawn up (paper/tunnel/therion)
- [Now go to the current documentation for adding a cave]
You are not finished until all of these are done
Actually adding a cave (1)
- Survey data [as survex file] in loser/caves/<number>
- Simple caves go in:
- location, name, description in expoweb/1623/cave_data/
- Complex caves go in:
- expoweb/1623/<cavenumber>/
- [Now go to the current documentation for adding a cave]
Actually adding a cave (2)
- Notes go into physical survey folder.
- Scans go into expofiles/surveyscans/<year>/<year>-<folderID>
- QMs don't seem to have a plan. Needs standardising [now as ;QM in .svx files]
- Caves with no number are called <year>-<initials>-<nn> and go in file expoweb/cave_data/<kataster_area>-<year>-<initials>-<nn> [e.g. 1623-2018-dm-07.html]
- [Now go to the current documentation for adding a cave]
Troggle
Django-based.
URL list filter /cave, /expedition, /survey
Templates for pages
Processors to generate them
Parsers to read in from website files.
Database back end (mysql)
The files are the 'real' database.
Run databaseReset.py reset
Troggle updating
For a new year, edit expoweb/noinfo/folk.csv
This is the thing that 'creates' a new 'expedition' (not the logbook!)
Run databaseReset.py reset
to reparse everything.
|