CfP: 2017 German Stata Users Group Meeting

Stata-Banner

We would like to announce the 15th German Stata Users Group meeting to be held Friday, June 23, 2017 at Humboldt University Berlin, Grimm-Zentrum, Geschwister-Scholl-Straße 1, 10117 Berlin.

All Stata users, from Germany and elsewhere, or those interested in learning about Stata, are invited to attend.

Presentations are sought on topics that include the following:

  • User-written Stata programs
  • Case studies of research or teaching using Stata
  • Discussions of data management problems
  • Reviews of analytic issues
  • Surveys or critiques of Stata facilities in specific fields, etc.

The conference language will be English, due to the international nature of the meeting and the participation of non-German guest speakers.

Submission guidelines

If you are interested in presenting a paper, please submit an abstract by email to one of the scientific organizers (max 200 words). The deadline for submissions is March 1, 2017. Presentations should be 20 minutes or shorter.

Registration

Participants are asked to travel at their own expense. There will be a small conference fee to cover costs for refreshments and lunch. There will also be an optional informal meal at a restaurant in Berlin on Friday evening at additional cost.

You can enroll by contacting Christiane Senczek by email or by writing or phoning.

Christiane Senczek
Dittrich & Partner Consulting GmbH
Prinzenstr. 2
42697 Solingen
Tel: +49 (0) 212 2 60 66-0
Email: christiane.senczek@dpc.de

The final program, cost, and venue will be circulated in April 2017.

Cost

  • Meeting only: 45 € (students 35 €)
  • Workshop only: 65 €
  • Workshop and Meeting: 85 €

Organizers

  • Scientific Organizers
  • Logistics Organizer
    • Dittrich & Partner Consulting GmbH (dpc.de), the distributor of Stata in several countries, including Germany, the Netherlands, Austria, the Czech Republic, and Hungary.

Data in the Middle: The common language of research (#iassist17)

IASSTS 2017 Logo
IASSTS 2017 Logo

The 43rd annual conference of the International Association for Social Science Information Services and Technology (IASSIST) will be held in Lawrence, Kansas from May 23-26, 2017.

Many issues around data (sources, strategies, and tools) are similar across disciplines. While IASSIST has its roots in social science data, it has also welcomed discussions over the years of other disciplines’ issues as they relate to data, data management, and support of users. So again this year, in line with this tradition, we are arranging a conference that will benefit those who support researchers across all disciplines: social sciences, health and natural sciences, and humanities. Please join the international data community in Lawrence, KS, “in the middle” of the U.S., for insights and discussion on how data in all disciplines are found, shared, used, and managed. Join us and draw inspiration from this diverse gathering! 

We welcome submissions for papers, presentations, panels, posters, and pecha kuchas.

The full Call for Proposals, along with the link to the submission form, can be accessed on the conference website here: http://www.iassist17.dept.ku.edu/proposals/

Questions can be directed to the Program Chairs, Samantha Guss and Michele Hayslett, at iassist2017@gmail.com.

We are also accepting submissions for Pre-conference Workshops under a separate Call for Workshops, which can be accessed here: http://www.iassist17.dept.ku.edu/proposals/workshops/

Questions about workshops may be sent to the Workshop Coordinators, Jenny Muilenburg (jmuil@uw.edu) and Andy Rutkowski (arutkowski@library.ucla.edu).

Deadline for all submissions: 21 November 2016.

Notification of acceptance: February 2017.

Am DIW Berlin für das SOEP arbeiten

SOEP-Logo
SOEP-Logo

Die im DIW Berlin angesiedelte forschungsbasierte Infrastruktureinrichtung Sozio-oekonomisches Panel (SOEP) sucht für ihr Datenaufbereitungsteam ab 1. Januar 2017 eine(n)

MitarbeiterIn im Bereich der
sozial- und wirtschaftswissenschaftlichen Dateninfrastruktur

Aufgabengebiete

  • Mitarbeit bei der Durchführung und konzeptionellen Weiterentwicklung der laufenden Datenaufbereitung
  • Weiterführung sowie Koordination des nutzerfreundlichen Datenformats SOEPlong
  • Weiterentwicklung bzw. ggf. Neukonzeption der Generierung von SOEPlong
    Mitarbeit bei der inhaltlichen Nutzung und Erstellung von Metadaten (lfd. Aktualisierung,  Aufbereitung und Weiterentwicklung)
  • Datenschutzprüfungen von Programmoutput bei kontrollierter Datenfernverarbeitung (remote execution bzw. remote access)
  • Generierung von nutzerfreundlichen Variablen für sozial- und wirtschaftswissenschaftliche Anwendungen

Informationen zu Anforderungen, Rahmenbedingunen und Bewerbung.

ESRA 2017 in Lissabon: Abstracts bis 4. Dezember 2016 einreichen

ESRA2017-Logo
Banner der ESRA in Lissabon 2017

Die Konferenzen der European Survey Research Association (ESRA) bieten auch immer die Möglichkeit, Themen aus Forschungsdatenzentren zu diskutieren. Die 7. ESRA-Konferenz findet im Juli 2017 in Lissabon statt. Der Call for Papers läuft noch bis 4. Dezember 2016, es wurden 111 Sessions angenommen.

Sessions im Bereich “Data Management and Archiving” (Auswahl):

Sessions im Bereich “Survey Methodological Research” (Auswahl):

Das ist eine subjektive Auswahl von Sessions mit inhaltlichem Bezug zu diesem Blog. Sie wird gerne ergänzt.

Workshop „Datenaufbereitung und Dokumentation“ am 20./21. Februar 2017 in Nürnberg (korrigiert)

Das IAB/PASS richtet am 20./21. Februar 2017 den nächsten Workshop „Datenaufbereitung und Dokumentation“ in Nürnberg aus.

Fast schon traditionell findet der Workshop im Vorfeld des dann mittlerweile 11. Workshops der deutschsprachigen Panelsurveys statt, der direkt im Anschluss am 21. und 22. Februar 2017 geplant ist.

Informationen zu Programm, Anmeldung und Unterkunft für beide Veranstaltungen folgen.

(Update 17.10.2016: Das Datum wurde eben durch den Veranstalter korrigiert.)

Open-Science-Webinare im Oktober

Das Center for Open Science bietet im Oktober drei Webinare auf englisch an:

RWX: Foundations of DDI in Scientific Literature

DDI DIRECTIONS: Logo of the DDI Alliance's Newsletter

DDI DIRECTIONS: Logo of the DDI Alliance’s Newsletter

Im aktuellen Newsletter DDI DIRECTIONS der DDI Alliance, der insbesondere über die Mailingliste [DDI-users] versandt wird, ist erstmals die Kolumne Read-Write-Execute (RWX) erschienen, die auf wissenschaftliche Publikationen aus dem Bereich Metadaten hinweisen will. Die Kolumne könnte auch ein Beispiel sein für entsprechende Informationen aus der (deutschsprachigen) Forschungsdatenszene, Interessierte können sich gerne bei Knut Wenzig melden. Hier die Ankündigung und der Text aus dem Newsletter:

The DDI Community has produced a rich store of DDI and metadata-related publications over the last 20 years. Read-Write-Execute (RWX) will highlight some of these existing publications as well as new work as it is produced. This first column will feature some of the foundations of DDI in scientific literature. (Thanks to Achim Wackerow for his suggestions and to Kelly Chatain for editorial assistance.) At the same time, a bibliography of DDI articles, working papers, and presentations is being built and is available at Bibsonomy.org with easily reusable bibliographic metadata. This metadata will also be made available on the DDI Alliance website. Suggestions for papers and topics for RWX, or the bibliography, are appreciated and can be sent to: Knut Wenzig, kwenzig@diw.de.

In her paper “The DDI matures: 1997 to the Present”, Mary Vardigan (2013), the former Director of the DDI Alliance, presents a timeline of the conceptual and organizational development of DDI. The initial SGML Codebook Committee meeting occurred in 1995, but 1997 was the year of the first “instantiation in XML” (Vardigan 2013: 45). DDI started (in versions 1 and 2) to describe data sets by the codebook approach, which is still supported and widely in use. From version 3 on the scope was broadened to document the whole lifecycle of data, using data collection as a starting point, and finally enabling repurposing and reuse of DDI elements.[1] The paper ends with a list of high level design goals, referred to in “Developing a Model-Driven DDI Specification” (Participants in 2012 Dagstuhl Seminar on DDI Moving Forward, 2012), on which the next version of DDI is based.

The first reference listed in Mary Vardigan’s paper is “Providing Global Access to Distributed Data Through Metadata Standardisation – the Parallel Stories of NESSTAR and the DDI”, submitted by the Norwegian Social Science Data Services and prepared by Jostein Ryssevik (1999). The “relative distance between the end-users of a statistical material and the production process” (p. 2) was identified as the fundamental problem to be solved. As discovery systems were provided to address this problem, the need for metadata standards like DDI emerged. The authors recall that DDI used the new (at the time) XML language, and that the defined XML code could contain the description of the document itself, of the study, the file, the variables, and other study-related materials. Already in this early paper RDF (Resource Description Framework[2]) is described as an application “that provides the foundation for metadata interoperability across different resource description communities.” (p. 5) Using DDI as a language, the medium NESSTAR could deliver a great range of interconnected services and platforms. Even if the last release of NESSTAR is more than one year old, the ideas in the article – whether or not realized by the software – deserve to be revisited. Using the metaphoric antonym of Bazaars vs. Cathedrals, the same authors (Ryssevik 2000) conceptualize their vision of – even then! – metadata systems that cover the complete life-cycle.

The article “The Data Documentation Initiative”, by Grant Blank and Karsten Boye Rasmussen (2004), was published in Social Science Computer Review, one of the top ranked academic journals in the “Information Science & Library Science” category. The authors describe the requirements of data documentation in the social sciences, how DDI as an XML based standard can be used to store information presented in codebooks, and how “standardization creates new opportunities for software development to aid users.” (p. 314)

Today, after 20 years, we can read and reevaluate those ideas only because people took the time to write them down. In this sense contributing to the scientific inventories of knowledge should be understood as a best practice and an integral part of software development for the academic community.

[1] Version 3 is described in Vardigan, Heus, Thomas 2008.
[2] Vardigan (2013: 48) expects that RDF will be used in the upcoming version of DDI as a connection to the semantic web.

References

(also available at Bibsonomy)

Blank, G. & Rasmussen, K. B. (2004), ‘The Data Documentation Initiative: The Value and Significance of a Worldwide Standard’, Social Science Computer Review 22 (3), 307-318, doi:10.1177/0894439304263144.

Participants in 2012 Dagstuhl Seminar on DDI Moving Forward (2012), Developing a Model-Driven DDI Specification, DDI Working Paper Series (Other Topics) DDI Alliance, doi:10.3886/DDIWorkingPaper04.

Ryssevik, J. & The Norwegian Social Science Data Services (1999), Providing Global Access to Distributed Data through Metadata Standardisation–the Parallel Stories of Nesstar and the DDI, Conference of European Statistics, UN/ECE Work Session on Statistical Metadata (Geneva, Switzerland, 22-24 September 1999), Working Paper 10, http://www.unece.org/stats/documents/1999/09/metis/10.e.pdf.

Ryssevik, J. & The Norwegian Social Science Data Services (2000), Bazaar Style Metadata in the Age of the Web–An ‘Open Source’ Approach to Metadata Development, Conference of European Statistics, UN/ECE Work Session on Statistical Metadata (Washington D.C., United States, 28-30 November 2000), Working Paper 4, http://www.unece.org/fileadmin/DAM/stats/documents/2000/11/metis/4.e.pdf.

Vardigan, M.; Heus, P. & Thomas, W. (2008), ‘Data Documentation Initiative: Toward a Standard for the Social Sciences’, International Journal of Digital Curation 3 (1), 107-113, doi:10.2218/ijdc.v3i1.45.

Vardigan, M. (2013), ‘The DDI Matures: 1997 to the Present’, IASSIST Quarterly 37 (1–4), 45-50, http://www.iassistdata.org/sites/default/files/iq/iqvol371_4_vardigan.pdf

SMS versenden

Bisweilen ist es gewünscht, dass Passwörter nicht über E-Mail, sondern einen anderen Kanal verschickt werden. Idealerweise soll dann der Versand vom Rechner (und nicht etwa vom privaten Smartphone) angestoßen werden. Eine Übersicht über Anbieter gibt teltarif.de. Wir haben auch selbst noch etwas gesucht, und sind noch auf personal-voip.de und sipgate.de gestoßen. Die Kosten für Angebote mit eigener Absenderkennung belaufen sich auf maximal 10,2 cent pro Nachricht.

Letztendlich hat uns sloono.de mit diesen Features überzeugt:

  • weltweiter Versand und
  • frei wählbare Absenderkennung (in unserem Fall „SOEPhotline“) ohne Übertragung einer Mobilfunknummer.

Interessant ist darüber hinaus:

  • eine API, die den Versand etwa auch per E-Mail anstoßen lässt und
  • ein Windows-7-Widget, für dessen Installation keine Rechte benötigt werden.

Positiv zu berichten ist auch, dass der Support sofort reagiert und auf Anfrage die gewünschte Absenderkennung als Standard eingestellt hat, damit die zur Registrierung verwendete private Handynummer nicht versehentlich übermittelt wird. Die Weboberfläche schaltet etwa beim Login nicht automatisch auf https um, dafür muss man selbst sorgen.

 

„Cross-Cultural Survey Guidelines“ komplett überarbeitet

Survey Lifecycle from: Survey Research Center. (2016). Guidelines for Best Practice in Cross-Cultural Surveys. Ann Arbor, MI: Survey Research Center, Institute for Social Research, University of Michigan. Retrieved September, 12, 2016, from http://www.ccsg.isr.umich.edu/
Survey Lifecycle from: Survey Research Center. (2016). Guidelines for Best Practice in Cross-Cultural Surveys. Ann Arbor, MI: Survey Research Center, Institute for Social Research, University of Michigan. Retrieved September, 12, 2016, from http://www.ccsg.isr.umich.edu/

Weil der ganze Lifecycle abgedeckt ist, werden die komplett überarbeiteten „Cross-Cultural Survey Guidelines“ auch für Datenproduzent_innen interessant sein.

Dies wird insbesondere auf die Kapitel zu Paradaten, Datenharmonisierung, -aufbereitung und -vertrieb zutreffen, aber auch das umfangreiche Glossar dürfte in vielen Fällen hilfreich sein.

Hier der englische Ankündigungstext:

New and Expanded Cross-Cultural Survey Guidelines

First published in 2008, the Cross-Cultural Survey Guidelines have recently undergone a significant update and expansion (Beta release: July 2016). The new edition includes over 800 pages of content with major updates and the expansion of all existing chapters, as well as the addition of new chapters on study design, study management, paradata, and statistical analysis. More than 70 professionals from 35 organizations contributed to this effort.  The senior editor was Tom W. Smith of NORC at the University of Chicago.  See: http://ccsg.isr.umich.edu/index.php/about-us/contributions for a complete list of contributors.

 The Cross-Cultural Survey Guidelines were developed to provide information on best practices across the survey lifecycle in a world in which the number and scope of studies covering multiple cultures, languages, nations, or regions has increased significantly. They were the product of an initiative of the International Workshop on Comparative Survey Design and Implementation (http://www.csdiworkshop.org/). The initiative was led by Beth-Ellen Pennell, currently the director of international survey operations at the Survey Research Center, Institute for Social Research at the University of Michigan.  

The aim of the initiative was to develop and promote internationally recognized guidelines that highlight best practice for the conduct of comparative survey research across cultures and countries. The guidelines address the gap in the existing literature on the details of implementing surveys that are specifically designed for comparative research, including what aspects should be standardized and when local adaptation is appropriate. The intended audience for the guidelines includes researchers and survey practitioners planning or engaged in what are increasingly referred to as multinational, multiregional, or multicultural (3MC) surveys, although much of the material is also relevant for single country surveys.

The guidelines cover all aspects of the survey lifecycle and include the following chapters: Study Design and Organizational Structure; Study Management;  Tenders, Bids and Contracts; Sample Design; Questionnaire Design; Adaptation; Translation; Instrument Technical Design; Interviewer Recruitment, Selection, and Training; Pretesting; Data Collection; Paradata and Other Auxiliary Data; Data Harmonization; Data Processing and Statistical Adjustment; Data Dissemination; Survey Quality  and Ethical Considerations. The guidelines can be found at: http://ccsg.isr.umich.edu.  We welcome feedback and suggestions.