Refine
Year of publication
Document Type
- Conference Proceeding (52) (remove)
Has Fulltext
- yes (52)
Is part of the Bibliography
- no (52)
Keywords
- Text Mining (5)
- Concreteness (4)
- Semantik (4)
- Ausbildung (3)
- Bibliothek (3)
- German (3)
- Information Retrieval (3)
- Informationsmanagement (3)
- Klassifikation (3)
- Bibliothekswesen (2)
- Contract Analysis (2)
- Deutsch (2)
- Digitalisierung (2)
- Disambiguation (2)
- Distributional Semantics (2)
- E-Learning (2)
- Grader (2)
- Graja (2)
- Konkretum <Linguistik> (2)
- Kulturerbe (2)
- Machine Learning (2)
- Modellversuch BID (2)
- Open Access (2)
- Programmieraufgabe (2)
- Rechtswissenschaften (2)
- Sachtext (2)
- Sprachnorm (2)
- Vergleich (2)
- Vertrag (2)
- Wikibase (2)
- Wikidata (2)
- Ähnlichkeit (2)
- 3D data (1)
- Abbreviations (1)
- Abkürzung (1)
- Acronyms (1)
- Akronym (1)
- Algorithmus (1)
- Ambiguität (1)
- Annotation (1)
- Autobewerter (1)
- Automatische Klassifikation (1)
- Automatische Sprachanalyse (1)
- Automatisierte Programmbewertung (1)
- Azyklischer gerichteter Graph (1)
- Benutzererlebnis (1)
- Bewertungsaspekt (1)
- Bewertungsmaßstab (1)
- Bibliothekar (1)
- Bilderkennung (1)
- Bildersprache (1)
- Bildersuchmaschine (1)
- Bildmaterial (1)
- Bildverarbeitung (1)
- Book of Abstract (1)
- COVID-19 (1)
- Citizens (1)
- Classification (1)
- Codierung (1)
- Computerlinguistik (1)
- Constructive Alignment (1)
- Corpus construction (1)
- Data Science (1)
- Data Sharing (1)
- Data-Warehouse-Konzept (1)
- Datenaufbereitung (1)
- Decision Support Systems, Clinical (1)
- Deep Convolutional Networks (1)
- Dewey-Dezimalklassifikation (1)
- Didactic (1)
- Digital Wellbeing (1)
- Digitalization (1)
- Digitization (1)
- Disambiguierung (1)
- Dokumentanalyse (1)
- E - Assessment (1)
- FHIR (1)
- Fachsprache (1)
- Fassung (1)
- Feature and Text Extraction (1)
- Figurative Language (1)
- Focus Group (1)
- Formelhafte Textabschnitte (1)
- Forschungsdaten (1)
- GECCO: German Corona Consensus Data Set (1)
- Gesundheitsfürsorge (1)
- Gesundheitsinformationssystem (1)
- Graph-based Text Representations (1)
- Grappa (1)
- Gruppeninterview (1)
- Health IT (1)
- Health Information Interoperability (1)
- Hochschule (1)
- Home Care (1)
- Hybrid Conference (1)
- Image Recognition (1)
- Image Retrieval (1)
- Imagery (1)
- Images (1)
- Information Dissemination (1)
- Information Extraction (1)
- Information Management (1)
- Information Science (1)
- Interoperabilität (1)
- Java <Programmiersprache> (1)
- Keyword Extraction (1)
- Knowledge Maps (1)
- Kompakkt (1)
- Kompetenz (1)
- Korpus <Linguistik> (1)
- Krankenhaus (1)
- Krankenunterlagen (1)
- LIG (1)
- LOINC (1)
- Latent Semantic Analysis (1)
- Layout Detection (1)
- Legal Documents (1)
- Legal Writings (1)
- Legende <Bild> (1)
- Lemmatization (1)
- Lernmotivation (1)
- Lexical Semantics (1)
- Linear Indexed Grammars (1)
- Linked Data (1)
- Linked Open Data (1)
- Liver Transplantation (1)
- Markov Models (1)
- Maschinelles Lernen (1)
- Media Didactic Concept (1)
- Medical Coding (1)
- Mediendidaktik (1)
- Medizin (1)
- Medizinische Bibliothek (1)
- Middleware (1)
- Motivation (1)
- NFDI (1)
- NFDI4Culture – Konsortium für Forschungsdaten materieller und immaterieller Kulturgüter (1)
- NLP (1)
- Nierentransplantation (1)
- Notation <Klassifikation> (1)
- Open Repositories (1)
- Open Science (1)
- Open Source (1)
- OpenRefine (1)
- PDF <Dateiformat> (1)
- PDF Document Analysis (1)
- POS Tagging (1)
- Paraphrase (1)
- Paraphrase Similarity (1)
- Patient empowerment (1)
- Phraseologie (1)
- Physics (1)
- Physik (1)
- Plugin (1)
- ProFormA-Aufgabenformat (1)
- Qualifikation (1)
- Quality Control (1)
- Qualitätskontrolle (1)
- Rechtsdokumente (1)
- Reduction of Complexity (1)
- Regional Development (1)
- Regional Innovation Systems (1)
- Regional Policy (1)
- Repository <Informatik> (1)
- Schlagwortkatalog (1)
- Schlagwortnormdatei (1)
- Scientific image search (1)
- Selbstgesteuertes Lernen (1)
- Self-directed Learning (1)
- Semantics (1)
- Semantisches Datenmodell (1)
- Similarity Measures (1)
- Spezialbibliothekar (1)
- Standardised formulation (1)
- Standardisierung (1)
- Statistical Methods (1)
- Statistische Methoden (1)
- Structural Analysis (1)
- Systems Librarian, Data Librarian, Job advertisement analysis, Job profiles, New competencies (1)
- Terminologie (1)
- Terminology (1)
- Territorial Intelligence (1)
- Text Similarity (1)
- Text annotation (1)
- Textbooks (1)
- Thesaurus (1)
- Title Matching (1)
- Transplantatabstoßung (1)
- Verbal Idioms (1)
- Versicherungsvertrag (1)
- Vertragsklausel (1)
- Wikimedia Commons (1)
- Wikipedia categories (1)
- Wissenschaftliche Bibliothek (1)
- Word Norms (1)
- Wort (1)
- XML (1)
- Zweiwortsatz (1)
- abstractness (1)
- concreteness (1)
- context vectors (1)
- cultural heritage (1)
- data warehouse (1)
- distributional semantics (1)
- e-Assessment (1)
- eLearning (1)
- education (1)
- fall prediction (1)
- fall prevention (1)
- fall risk (1)
- graft rejection (1)
- high-quality Learning Formats (1)
- image processing (1)
- information extraction (1)
- interoperability (1)
- kidney transplant (1)
- library and information science (1)
- linked data (1)
- openEHR (1)
- research data management (1)
- research information (1)
- sensor-based assessment (1)
- supervised machine learning (1)
- thesauri (1)
- wearable sensors (1)
- web crawling (1)
- word embedding space (1)
- Öffentliche Bibliothek (1)
- Überwachtes Lernen (1)
Institute
- Fakultät III - Medien, Information und Design (52) (remove)
All of us are aware of the changes in the information field during the last years. We all see the paradigm shift coming up and have some idea how it will challenge our profession in the future. But how the road to excellence - in education of information specialists in the future - will look like? There are different models (new and old ones) for reorganising the structure of education: * Integration * Specialisation * Step-by step-model * Modul System * Network System / Combination model The paper will present the actual level of discussion on building up a new curriculum at the Department of Information and Communication (IK) at the FH Hannover. Based on the mission statement of the department »Education of information professionals is a part of the dynamic evolution of knowledge society« the direction of change and the main goals will be presented. The different reorganisation models will be explained with its objectives, opportunities and forms of implementation. Some examples will show the ideas and tools for a first draft of a reconstruction plan to become fit for the future. This talk has been held at the German-Dutch University Conference »Information Specialists for the 21st Century« at the Fachhochschule Hannover - University of Applied Sciences, Department of Information and Communication, October 14 -15, 1999 in Hannover, Germany.
Data and Information Science: Book of Abstracts at BOBCATSSS 2022 Hybrid Conference, 23rd - 25th of May 2022, Debrecen.
This year marks the 30th anniversary of the BOBCATSSS. The BOBCATSSS is an international, annual symposium designed for librarians and information professionals in a rapidly changing environment. Over the past 30 years, the conference has included exciting topics, great venues, interested guests and engaging presenters.
This year we would like to introduce the topics of the many papers presented in the Book of Abstracts for the first time in presence at the University of Debrecen and hybrid. The Book of Abstracts provides an overview of all presentations given at BOBCATSSS. Presentations are listed in alphabetical order by title and include speeches, Pecha Kuchas, posters and workshops.
The theme of BOBCATSSS is Data and Information Science. Data and information are the basis for decisions and processes in business, politics and science. Particularly important in the current era of digital transformation. This is exactly where this year's subthemes come in. They deal with data science, openness as well as institutional roles.
Lemmatization is a central task in many NLP applications. Despite this importance, the number of (freely) available and easy to use tools for German is very limited. To fill this gap, we developed a simple lemmatizer that can be trained on any lemmatized corpus. For a full form word the tagger tries to find the sequence of morphemes that is most likely to generate that word. From this sequence of tags we can easily derive the stem, the lemma and the part of speech (PoS) of the word. We show (i) that the quality of this approach is comparable to state of the art methods and (ii) that we can improve the results of Part-of-Speech (PoS) tagging when we include the morphological analysis of each word.
Automatic classification of scientific records using the German Subject Heading Authority File (SWD)
(2012)
The following paper deals with an automatic text classification method which does not require training documents. For this method the German Subject Heading Authority File (SWD), provided by the linked data service of the German National Library is used. Recently the SWD was enriched with notations of the Dewey Decimal Classification (DDC). In consequence it became possible to utilize the subject headings as textual representations for the notations of the DDC. Basically, we we derive the classification of a text from the classification of the words in the text given by the thesaurus. The method was tested by classifying 3826 OAI-Records from 7 different repositories. Mean reciprocal rank and recall were chosen as evaluation measure. Direct comparison to a machine learning method has shown that this method is definitely competitive. Thus we can conclude that the enriched version of the SWD provides high quality information with a broad coverage for classification of German scientific articles.
Das ProFormA-Aufgabenformat wurde eingeführt, um den Austausch von Programmieraufgaben zwischen beliebigen Autobewertern (Grader) zu ermöglichen. Ein Autobewerter führt im ProFormA-Aufgabenformat spezifizierte „Tests“ sequentiell aus, um ein vom Studierenden eingereichtes Programm zu prüfen. Für die Strukturierung und Darstellung der Testergebnisse existiert derzeit kein graderübergreifender Standard. Wir schlagen eine Erweiterung des ProFormA-Aufgabenformats um eine Hierarchie von Bewertungsaspekten vor, die nach didaktischen Aspekten gruppiert ist und entsprechende Testausführungen referenziert. Die Erweiterung wurde in Graja umgesetzt, einem Autobewerter für Java-Programme. Je nach gewünschter Detaillierung der Bewertungsaspekte sind Testausführungen in Teilausführungen aufzubrechen. Wir illustrieren unseren Vorschlag mit den Testwerkzeugen Compiler, dynamischer Softwaretest, statische Analyse sowie unter Einsatz menschlicher Bewerter.
Regional Innovation Systems describe the relations between actors, structures and infrastructures in a region in order to stimulate innovation and regional development. For these systems the collection and organization of information is crucial. In the present paper we investigate the possibilities to extract information from websites of companies. First we describe regional innovation systems and the information types that are necessary to create them. Then we discuss the possibilities of text mining and keyword extraction techniques to extract this information from company websites. Finally, we describe a small scale experiment in which keywords related to economic sectors and commodities are extracted from the websites of over 200 companies. This experiment shows what the main challenges are for information extraction from websites for regional innovation systems.
The amount of papers published yearly increases since decades. Libraries need to make these resources accessible and available with classification being an important aspect and part of this process. This paper analyzes prerequisites and possibilities of automatic classification of medical literature. We explain the selection, preprocessing and analysis of data consisting of catalogue datasets from the library of the Hanover Medical School, Lower Saxony, Germany. In the present study, 19,348 documents, represented by notations of library classification systems such as e.g. the Dewey Decimal Classification (DDC), were classified into 514 different classes from the National Library of Medicine (NLM) classification system. The algorithm used was k-nearest-neighbours (kNN). A correct classification rate of 55.7% could be achieved. To the best of our knowledge, this is not only the first research conducted towards the use of the NLM classification in automatic classification but also the first approach that exclusively considers already assigned notations from other
classification systems for this purpose.
The CogALex-V Shared Task provides two datasets that consists of pairs of words along with a classification of their semantic relation. The dataset for the first task distinguishes only between related and unrelated, while the second data set distinguishes several types of semantic relations. A number of recent papers propose to construct a feature vector that represents a pair of words by applying a pairwise simple operation to all elements of the feature vector. Subsequently, the pairs can be classified by training any classification algorithm on these vectors. In the present paper we apply this method to the provided datasets. We see that the results are not better than from the given simple baseline. We conclude that the results of the investigated method are strongly depended on the type of data to which it is applied.
A new FOSS (free and open source software) toolchain and associated workflow is being developed in the context of NFDI4Culture, a German consortium of research- and cultural heritage institutions working towards a shared infrastructure for research data that meets the needs of 21st century data creators, maintainers and end users across the broad spectrum of the digital libraries and archives field, and the digital humanities. This short paper and demo present how the integrated toolchain connects: 1) OpenRefine - for data reconciliation and batch upload; 2) Wikibase - for linked open data (LOD) storage; and 3) Kompakkt - for rendering and annotating 3D models. The presentation is aimed at librarians, digital curators and data managers interested in learning how to manage research datasets containing 3D media, and how to make them available within an open data environment with 3D-rendering and collaborative annotation features.