Working with semi-structured or malformed data requires skills that cross disciplines. Fortunately there are tools that make cleaning and standardizing data easier. This workshop's goals are to introduce participants to OpenRefine, cover basic strategies for its use, and perform exercises to familiarize them with its capabilities.
Participants will complete guided exercises where they clean and sort sample data. Participants are expected to have a current Java SE Runtime Environment and an installation of the most recent version of OpenRefine. The Chrome browser, or its open-source sibling Chromium, are needed for full compatibility with OpenRefine's user interface.
Metadata is the driving force behind digital file management, but not everyone understands how embedded metadata can be, or should be utilized to help with digital file management. Learn how you can use file-naming conventions and specific tools such as Phil Harvey’s EXIFTool or Python scripts to edit, apply, and manage metadata to large batches of files to facilitate asset discovery down the road. We’ll also take a look at how you can read and write embedded metadata to digital files. These embedded metadata exercises can be used to apply to a number of different goals such as:
And you’ll also learn some tips and tricks for when you might want to use embedded metadata, and other cases where using embedded metadata is not the standard practice (such as on ecommerce websites.)
To use digital media effectively in both research and instruction, you need to go beyond just the playback of media files. You need to be able to stream the media, divide that stream into different segments, provide descriptive analysis of each segment, order, re-order and compare different segments from the same or different streams and create web sites that can show the result of your analysis. In this workshop, I will use Omeka and several plugins I have developed for working with digital media, to show the potential of video streaming, segmentation and descriptive analysis for research and instruction.
Omeka S represents a complete rewrite of Omeka Classic (aka the Omeka 2.x series), adhering to our fundamental principles of encouraging use of metadata standards, easy web publishing, and sharing cultural history. New objectives in Omeka S include multisite functionality and increased interaction with other systems. This workshop will compare and contrast Omeka S with Omeka Classic to highlight our emphasis on 1) modern metadata standards, 2) interoperability with other systems including Linked Open Data, 3) use of modern web standards, and 4) web publishing to meet the goals medium- to large-sized institutions.
In this workshop we will walk through Omeka S Item creation, with emphasis on LoD principles. We will also look at the features of Omeka S that ease metadata input and facilitate project-defined usage and workflows. In accordance with our commitment to interoperability, we will describe how the API for Omeka S can be deployed for data exchange and sharing between many systems. We will also describe how Omeka S promotes multiple site creation from one installation, in the interest of easy publishing with many objects in many contexts, and simplifying the work of IT departments.
Participants in the workshop will learn:
We all face failure in our professional lives, but no one likes to talk about it. Our relationship with failure frequently comes hand in hand with embarrassment, fear, and taboo. But failure has intrinsic value and is an essential step on the path to professional success. And since it's inevitable, we ought to learn how to face failure, how to talk about it as professionals, and how to grow from it. Fail4Lib, now in its 6th year, is the perennial Code4Lib preconference dedicated to discussing and coming to terms with the failures that we all encounter in our professional lives. It is a safe space for us to explore failure, to talk about our own experiences with failure, and to encourage enlightened risk taking. The goal of Fail4Lib is for participants -- and their organizations -- to get better at failing gracefully, so that when we do fail, we do so in a way that moves us forward. This half-day preconference will consist of case studies, round-table discussions, and, for those interested in sharing, lightning talks on failures we've dealt with in our own work.
Have you been curious about static website generators? Have you been wondering who “Jekyll” and “Hugo” are? Then this workshop is for you!
In this hands-on workshop, we’ll start by exploring static website generators, their components, some of the different options available, and their benefits and disadvantages. Then, we’ll work on making our own sites, and for those that would like to, get them online with GitHub pages. Familiarity with HTML, git, and command line basics will be helpful but are not required.
Introduction to transcription with Wikisource. Introduction to wikicode and Visual Editor (WYSIWYG) in Wikisource. Comparison with other transcription efforts. Practice editing with experienced Wikimedian tutors.
This workshop will do a deep dive into approaches and recommend best practices for customizing Blacklight applications. We will discuss a range of topics, including styling and theming, customizing discovery experiences, and working with Solr.
Spotlight is an open source application that extends the digital library ecosystem by providing a means for institutions to reuse digital content in easy-to-produce, attractive, and scholarly-oriented websites. Librarians, curators, and other content experts can build Spotlight exhibits to showcase digital collections using a self-service workflow for selection, arrangement, curation, and presentation.
This workshop will introduce the main features of Spotlight and present examples of Spotlight-built exhibits from the community of adopters. We'll also describe the technical requirements for adopting Spotlight and highlight the potential to customize and extend Spotlight's capabilities for their own needs while contributing to its growth as an open source project.
User research is often focused on measures of the usability of online spaces. We look at search traffic, run card sorting and usability testing activities, and track how users navigate our spaces. Those results inform design decisions through the lens of information architecture. This is important, but doesn't encompass everything a user needs in a space.
This workshop will focus on the other component of user experience design and user research: how to create spaces where users feel safe. Users bring their anxieties and stressors with them to our online spaces, but informed design choices can help to ameliorate that stress. This will ultimately lead to a more positive interaction between your institution and your users.
The presenters will discuss the theory behind empathetic design, delve deeply into using ethnographic research methods - including an opportunity for attendees to practice those ethnographic skills with student participants - and finish with the practical application of these results to ongoing and future projects.
At previous Code4Lib conferences, there have been wonderful workshops showcasing a variety of tools and methods useful in webscraping. However, every tool has its limitations. In this workshop, we aim to demonstrate some of these limitations via a case study approach in order to highlight the need for a robust toolkit of scraping tools and methodologies when collecting for archival and research purposes. Then, we use practical, hands-on demonstrations to assist workshop participants in developing their own robust toolkits for scraping the web.
Impostor syndrome is a drag. You think you aren’t good enough to do what you do, even if there’s a wealth of information refuting that. Maybe you feel lost, alone, and afraid while doing specific parts of your job. If this is a thing that feels familiar to you, come to this workshop! We’ll get some group therapy in, I’ll tell you some sweet (read: terrifying) stories from my own professional experience, and I’ll give you some great resources for battling that “I should be quiet in the corner” mentality. There will be a handout.
Am I qualified to lead this workshop? Probably not (I struggle with this every day), but I also think there should be something for people like us at every professional conference.
In this hands-on workshop, you will get introduced to the Python programming language through a fun project you can show off. With a starter kit of Python code provided by workshop leaders, you’ll customize your Twitter bot and see it send tweets out into the world. Why learn Python? It’s a great “gateway” language to programming. It’s commonly used, very human-readable, and especially suited to text processing tasks you may come across in library work. Why bots? Bots work behind the scenes of the web. They can tell us the weather forecast when we ask (hey Siri), they systematically fix broken links on Wikipedia, they summarize financial data (Forbes), and on Twitter, whimsical bots can add a little bit of fun to the Internet. (Our favorites include @JustToSayBot, @TwoHeadlines, and @big_ben_clock.) By the end of this workshop, you will have a functional bot of your very own and an elementary understanding of Python.
Open to everyone who has an interest in learning about programming. No prior experience needed.
Plug-ins provide a simple mechanism to customize ArchivesSpace without changing the core codebase. They can be used to extend or override built-in functionality and the look and feel of the application. Through examples and hands-on activities, this workshop will introduce the concept of plug-ins, demystify the process of building them, and provide participants with tips and tools for building their own.
Tech workshops pose two unique problems: finding skilled instructors for that content, and instructing that content well. Library hosted workshops are often a primary educational resource for solo learners, and many librarians utilize these workshops as a primary outreach platform. Tackling these two issues together often makes the most sense for our limited resources. Whether a programming language or software tool, learning tech to teach tech can be one of the best motivations for learning that tech skill or tool, but equally important is to learn how to teach and present tech well.
This hands-on workshop will guide participants through developing their own learning plan, reviewing essential pedagogy for teaching tech, and crafting a workshop of their choice. Each participant will leave with an actionable learning schedule, a prioritized list of resources to investigate, and an outline of a workshop they would like to teach.
The Internet of Things is a rising trend in library research. IoT sensors can be used for space assessment, service design, and environmental monitoring. IoT tools create lots of data that can be overwhelming and hard to interpret. Tableau Public is a data visualization tool that allows you to explore this information quickly and intuitively to find new insights. This full-day workshop will teach you the basics of building your own own IoT sensor using a Raspberry Pi in order to gather, manipulate, and visualize your data. All are welcome, but some familiarity with Python is recommended.
This workshop will focus on understanding and experiencing the interaction models defined by the web-standards that are contained in the Fedora Repository API. Specifically:
We will play with historic newspaper data by working through some guided tasks using the Chronicling America API.
This session will introduce two major types of virtualization, virtual machines using tools like VirtualBox and Vagrant, and containers using Docker. The relative strengths and drawbacks of the two approaches will be discussed along with plenty of hands-on time. Though geared towards integrating these tools into a development workflow, the workshop should be useful for anyone interested in creating stable and reproducible computing environments, and examples will focus on library-specific tools like Archivematica and EZPaarse. With virtualization taking a lot of the pain out of installing and distributing software, alleviating many cross-platform issues, and becoming increasingly common in library and industry practices, now is a great time to get your feet wet.
Social media data represents a tremendous opportunity for memory institutions of all kinds, be they large academic research libraries, or small community archives. Researchers from a broad swath of disciplines have a great deal of interest in working with social media content, but they often lack access to datasets or the technical skills needed to create them. Further, it is clear that social media is already a crucial part of the historical record in areas ranging from events your local community to national elections. But attempts to build archives of social media data are largely nascent. This workshop will be both an introduction to collecting data from the APIs of social media platforms, as well as a discussion of the roles of libraries and archives in that collecting.
Assuming no prior experience, the workshop will begin with an explanation of how APIs operate. We will then focus specifically on the Twitter API, as Twitter is of significant interest to researchers and hosts an important segment of discourse. Through a combination of hands-on and demos, we will gain experience with a number of tools that support collecting social media data (e.g., Twarc, Social Feed Manager, DocNow, Twurl, and TAGS), as well as tools that enable sharing social media datasets (e.g., Hydrator, TweetSets, and the Tweet ID Catalog).
The workshop will then turn to a discussion of how to build a successful program enabling social media collecting at your institution. This might cover a variety of topics including outreach to campus researchers, collection development strategies, the relationship between social media archiving and web archiving, and how to get involved with the social media archiving community. This discussion will be framed by a focus on ethical considerations of social media data, including privacy and responsible data sharing.
Time permitting, we will provide a sampling of some approaches to social media data analysis, including Twarc Utils and Jupyter Notebooks.
In this hands-on workshop, we’ll lead you through a series of information gathering and design thinking activities that will teach you methods to explore any problem, then ideate and refine your way to a solution. You’ll learn the user research tools needed to root an investigation in the perspective and needs of a target user. You'll learn how to conduct structured group-brainstorming that works for introverts and extroverts alike and generates dozens of ideas in no time. Last, you'll learn how to quickly refine and select an approach using rapid prototyping and user feedback. You and your team will leave with the tools needed to run your own innovation project, whether it be a new internal service, an externally-facing product or even a funded planning grant.
This hands-on workshop details the steps involved in successfully integrating Alma, Summon and OS discovery systems including Blacklight and VUFind. This will be a technical workshop--a deep dive with a review of the Alma and Summon APIs used in production today, an emphasis on obstacles faced during development, best practice documentation, and ongoing maintenance. Most importantly, the technical details of the integrations will be accompanied by the strategic decisions faced by libraries during the project. Come learn more about existing open source projects and how to develop within the open Ex Libris framework.
Library-generated data related to space utilization, technology utilization, and other primary services can be a main source of insight into the current state of operations. However, if trapped in spreadsheets and on server logs, the data is difficult to evaluate, and often remains unevaluated. Web-based data dashboards provide a platform independent way of incorporating this data into a near realtime visual summary.
In this hands on workshop we will utilize the web framework Angular and the visualization library D3.js to walk participants through the creation of a sample data dashboard application. Benefits of using the Angular framework include reusable components, two-way data binding, and access to command line tools which simplify code generation. D3.js provides a feature-rich library for crafting personalized and modular visualizations.
This is an introductory workshop to Solr, the fast and open source search platform that powers a lot of library products. This workshop is geared to anyone that has never used Solr or uses it but has not looked under the hood to see how it can be configured or explored some of the features that Solr offers out of the box.
We'll start the workshop with a quick review on how Solr stores data and the process that it goes when a search is submitted.
We'll then go through a tour of the main features in Solr:
While practical, this workshop does not require you to be a coder. The idea is for you to learn the concepts, features, and how they work. If you choose, we'll show you how to install Solr on your machine to experiment with the concepts as we go. But if you prefer not to install Solr (of updating configuration files is not your thing) you can still follow along and learn with us.
While MARC still remains the authoritative data source for library catalogs, projects like the Library of Congress's Marc2Bibframe2 and BIBCAT are among the first tools in extracting data out of MARC records into RDF BIBFRAME triples. The conversion process produces BIBFRAME RDF data that must be cleaned, sorted, and linked to the web. This working session provides a hands-on experience in manipulating linked-data using the BIBFRAME RDF vocabulary in building cataloging and bibliographic applications. Other topics include mapping BIBFRAME to Schema.org for search engine indexing and the challenges in the conversion process, such as resolving locally constructed entities to external IRIs. We will also cover the uses of BIBCAT and RDF Maps to extract data from non-RDF sources like MODS XML, CSV, JSON feeds, and SPARQL endpoints to BIBFRAME and Schema.org triples.
Libraries must ensure that users of all abilities can successfully use the technologies we provide. Despite the many ethical and legal motivations, not all of our technologies meet accessibility standards. Ultimately, the responsibility for making technologies accessible falls to the developers and vendors, but it does not begin with them. Advocacy from library staff of all duties is crucial for ensuring that disability access is a priority in library technology.
This workshop provides a foundation of skills and knowledge for becoming an accessibility advocate for any library worker. Background on disability types, assistive technologies, standards, and laws will be provided. Basic technology accessibility concepts will be covered, including simple tests and questions for working with vendors. Depending on the interests and background of workshop attendees, additional topics may be explored such as document formats, metadata, or testing practices.
No technical expertise is required of participants, although some discussions will get into HTML and other semantic markup languages.
This Mapping Veronese & Muses Workshop is based on leveraging Esri Story Mapping software and publicly available museum data. The workshop teaches simple ways to research and identify museum and art data on the internet; create a simple map; and create a story of images, history, and text that can be shared online. The software empowers curators and historians to develop a story with maps, texts, and imagery to complement online discussions, presentations, and tours. At the end of the workshop, participants will have a simple story map utilizing Esri that they can further build upon post-workshop.
This is an introductory session on Apache Spark, a framework for large-scale data processing. We will introduce high level concepts around Spark, including how Spark execution works and it’s relationship to the other technologies for working with Big Data. Following this introduction to the theory and background, we will walk workshop participants through hands-on usage of spark-shell, Zeppelin notebooks, and Spark SQL for processing library data. The workshop will wrap up with use cases and demos for leveraging Spark within cultural heritage institutions and information organizations, connecting the building blocks learned to current projects in the real world.
What can you do with apps on a platform? FOLIO is a platform for developing services geared towards libraries. With built-in support for handling patron data and a multitude of metadata formats, FOLIO provides the foundation to build new tools and integrate existing services into a cohesive whole.
This year’s Code4Lib marks the second anniversary of Sebastian Hammer’s description of the platform concept to the community, 18 months since the code repositories were opened, and a year since the FOLIO curriculum invited developers to test drive FOLIO concepts. The foundation is now broader and deeper. Participants will learn how to build and integrate new RESTful services on the back end as well as how to make use of the React- and Redux-based user interface components.
An introduction to the mechanics of ISLE, including a demo of new systems creation and existing systems maintenance processes. ISLE separates an institution's customizations from core code, and moves that core code into containers that are easily updated, simplifying and largely automating the process of installation and updates/maintenance of Islandora. ISLE also bundles together the best shared modules into a common, production-ready and security-hardened platform.
The International Image Interoperability Framework (IIIF) is a set of technical specifications built around shared challenges in cultural heritage access. This technical workshop will provide an overview of the IIIF specifications and hands-on exercises to gain a deeper understanding of the current landscape of tools and concepts.
Participants will explore the core Image and Presentation APIs, take a close look at various IIIF viewers, and get a brief introduction to the Content Search and Authentication APIs. Participants will also have an opportunity to discuss their institution's use cases and appropriate solutions, see some examples of use beyond the standard image viewers/tools, and receive pointers for how to go further and engage with the community.