Notice: On April 23, 2014, Statalist moved from an email list to a forum, based at statalist.org.
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
st: 2013 German Stata Users Group Meeting -- Program
From
Ulrich Kohler <[email protected]>
To
"[email protected]" <[email protected]>
Subject
st: 2013 German Stata Users Group Meeting -- Program
Date
Wed, 10 Apr 2013 10:08:29 +0200
Dear Stata-Users,
The 11th German Stata Users Group Meeting will be held at the the
University of Potsdam on Friday, June 7 2013. Everybody from anywhere
who is interested in using Stata is invited to attend this meeting. The
meeting will include presentations about causal models, general
statistics, and data management, both by researchers and by StataCorp
staff. The meeting will also include a "wishes and grumbles" session,
during which you may air your thoughts to Stata developers.
On the day before the conference, Ulrich Kohler, Co-author of the Book
"Data Analysis Using Stata" and author of several user written Stata
commands will hold a workshop on "Advanced Do-File programming and
introduction to Ado-file programs". Details about the workshop are given
below the program.
There is (at additional cost) the option of an informal meal at a
restaurant in Potsdam on Friday evening. Details about accommodations
and fees are given below.
The conference language will be English.
Also see: http://www.stata.com/meeting/germany13/
Date and Venue
-------------
June 7, 2013
University of Potsdam
Building Nr. 6, Room H06
Campus Griebnitzsee
August-Bebel-Str. 89
14482 Potsdam
http://www.uni-potsdam.de/wiso_dekanat/english
Costs
-----
Meeting only: 45 EUR (students 25 EUR)
Workshop on June 6: 65 EUR
Workshop and Conference: 85 EUR
Registration and accommodations
-------------------------------
Please travel at your own expense. You can enroll by contacting Anke
Mrosek ([email protected]) by email or by writing, phoning, or faxing to
Anke Mrosek
Dittrich & Partner Consulting GmbH
Prinzenstrasse 2
42697 Solingen Germany
Tel: +49 (0)212 260 6624
Fax: +49 (0)212 260 6666
Scientific Organizers
--------------------
The academic program of the meeting is being organized by J. Giesecke
(University of Bamberg) and and U. Kohler (University of Potsdam).
Logistics organizers
--------------------
The logistics are being organized by Dittrich and Partner
(http://www.dpc.de), the distributor of Stata in several countries
including Germany, The Netherlands, Austria, Czech Republic, and Hungary.
Program
-------
8:30–9:00 Registration
9:00–9:15 Welcome
Ulrich Kohler, University of Potsdam
9:15–10:15 Creating complex tables for publication
John Luke Gallup, Portland State University
Abstract: Complex statistical tables often must be built up by parts
from the results of multiple Stata commands. I show the capabilities of
frmttable and outreg for creating complex tables, and even fully
formatted statistical appendices, for Word and TeX documents. Precise
formatting of these tables from within Stata has the same benefits as
writing do-files for statistics commands. They are reproducible and
reusable when the data change, saving the user time.
10:15–11:15 An expanded framework for mixed process modeling in Stata
David Roodman, Center for Global Development
Abstract: Roodman (Stata Journal, 2011) introduced the program cmp for
using maximum likelihood to fit multiequation combinations of
Gaussian-based models such as tobit, probit, ordered probit, multinomial
probit, interval censoring, and continuous linear. This presentation
describes substantial extensions to the framework and software: factor
variable support; the rank-ordered probit model; the ability to specify
precensoring truncation in most model types; hierarchical random effects
and coefficients that are potentially correlated across equations; the
ability to include the unobserved linear variables behind endogenous
variables—not just their observed, censored manifestations—on the right
side of other equations and, when so doing, the allowance for
simultaneity in the system of equations. Contrary to the title of
Roodman (2011), models no longer need be recursive or fully observed.
11:15–11:30 Coffee
11:30–11:45 Provide, Enrich and Make Accessible: Using Stata’s
Capabilities for Disseminating NEPS Scientific Use Data
Daniel Bela, National Educational Panel Study (NEPS), Data Center
University of Bamberg A
Abstract: The National Educational Panel Study (NEPS) is rising as one
of Germany’s major publisher of scientific use data for educational
research. Disseminating data from six panel cohorts makes not only
structured data editing but also documentation and user support a major
challenge. In order to accomplish this task, the NEPS Data Center has
implemented a sophisticated metadata system. It does not only allow the
structured documentation of the metadata of survey instruments and data
files. It also allows one to enrich the scientific use files with further
information, thus significantly easing access for data analyses. As a
result, NEPS provides bilingual dataset files (German and English) and
allows the user to instantly see, for instance, the exact wording of the
question leading to the data in a distinct variable without leaving the
dataset. To achieve this, structured metadata is attached to the data
using Stata’s characteristics functionality. To make handling additional
metadata even easier, the NEPS Data Center provides a package of
user-written programs, NEPStools, to data users. The presentation will
cover an introduction to the NEPS data preparation workflow, focusing on
the metadata system and its role in enriching the scientific use data by
using Stata’s capabilities. Afterward, NEPStools will be introduced.
11:45–12:00 newspell—Easy Management of Complex Spell Data
Hannes Neiss, German Institute for Economic Research (DIW) and Berlin
Graduate School of Social Sciences (BGSS)
Abstract: Biographical data gathered in surveys is often stored in spell
format, allowing for overlaps between spell states. This gives useful
information to researchers but leaves them with a very complex data
structure, which is not easy to handle. I present my work on the
ado-package newspell. It includes several subprograms for management of
complex spell data. Spell states can be merged, reducing the overall
number of spells. newspell allows a user to fill gaps with information
from spells before and after the gap, given a user-defined preference.
However, the two most important features of newspell are, first, the
ability to rank spells and cut off overlaps according to the rank order.
This is a necessary step before performing, for example, sequence
analysis on spell data. Second, newspell can combine overlapping spells
into new categories of spells, generating entirely new states. This is
useful for cleaning data, for analyzing simultaneity of states, or for
combining two spell datasets that have information on different kinds of
states (for example, labor market and marital status). newspell is
useful for users who are not familiar with complex spell data and have
little experience in Stata programming for data management. For
experienced users, it saves a lot of time and coding work.
12:00–12:30 Instrumental variables estimation using
heteroskedasticity-based instruments
Christopher F. Baum, Arthur Lewbel (Boston College) Mark E. Schaffer
(Heriot–Watt University, Edinburgh), Oleksandr Talavera (University of
Sheffield)
Abstract: In a 2012 article in the Journal of Business and Economic
Statistics, Arthur Lewbel presented the theory of allowing the
identification and estimation of “mismeasured and endogenous regressor
models” by exploiting heteroskedasticity. These models include linear
regression models customarily estimated with instrumental variables (IV)
or IV-GMM techniques. Lewbel’s method, under suitable conditions, can
provide instruments where no conventional instruments are available or
augment standard instruments to enable tests of overidentification in the
context of an exactly identified model. In this talk, I discuss the
rationale for Lewbel’s methodology and illustrate its implementation in
a variant of Baum, Schaffer, and Stillman’ sivreg2 routine, ivreg2h.
12:30–13:30 Lunch
13:30–14:00 Using simulation to inspect the performance of a test, in
particular tests of the parallel regressions assumption in ordered logit
and probit models
Maarten L. Buis, Social Science Research Center (WZB) and Richard
Williams, University of Notre Dame
Abstract: In this talk, we will show how to use simulations in Stata to
explore to what extent and under what circumstances a test is
problematic. We will illustrate this for a set of tests of the parallel
regression assumption in ordered logit and probit models: the Brant,
likelihood ratio, Wald, score, and Wolfe-Gould test of the parallel
regression assumption. A common impression is that these tests tend to
be too anti-conservative; that is, they tend to reject a true null
hypothesis too often. We will use simulations to try to quantify when
and to what extent this is the case. We will also use these simulations
to create a more robust bootstrap variation of the tests. The purpose of
this talk is twofold: first, we want to explore the performance of these
tests. For this purpose, we will present a new program, oparallel, that
implements all tests and their bootstrap variation. Second, we want to
give more general advice on how to use Stata to create simulations when
one has doubts about a certain test. For this purpose, we will present
the simpplot command, which can help to interpret the p-values returned
by such a simulation.
14:00–14:30 Fitting Complex Mixed Logit Models with Particular Focus on
Labor Supply Estimation
Max Löffler, Institute for the Study of Labor (IZA)
Abstract: When one estimates discrete choice models, the mixed logit
approach is commonly superior to simple conditional logit setups. Mixed
logit models not only allow the researcher to implement difficult random
components but also overcome the restrictive IIA assumption. Despite
these theoretical advantages, the estimation of mixed logit models
becomes cumbersome when the model’s complexity increases. Applied works
therefore often rely on rather simple empirical specifications because
this reduces the computational burden. I introduce the user-written
command lslogit, which fits complex mixed logit models using maximum
simulated likelihood methods. As lslogit is a d2-ML-evaluator written in
Mata, the estimation is rather efficient compared with other routines. It
allows the researcher to specify complicated structures of unobserved
heterogeneity and to choose from a set of frequently used functional
forms for the direct utility function—for example, Box-Cox
transformations, which are difficult to estimate in the context of logit
models. The particular focus of lslogit is on the estimation of labor
supply models in the discrete choice context; therefore, it facilitates
several computationally exhausting but standard tasks in this research
area. However, the command can be used in many other applications of
mixed logit models as well.
14:30–15:00 Simulated Multivariate Random Effects Probit Models for
Unbalanced Panels
Alexander Plum, Otto-von-Guericke University Magdeburg
Abstract: This paper develops an implementation method of a simulated
multivariate random-effects probit model for unbalanced panels,
illustrating it by using artificial data. By mdraws, generated Halton
draws are used to simulate multivariate normal probabilities with the
command mvnp(). The estimator can be easily adjusted (for example, to
allow for autocorrelated errors). Advantages of this simulated
estimation are high accuracy and lower computation time compared with
existing commands such as redpace.
15:00–15:15 Coffee
15:15–15:45 xsmle—A Command to Estimate Spatial Panel Models in Stata
Federico Belottia, Gordon Hughes, Andrea Piano Mortari CEIS, University
of Rome "Tor Vergata" and School of Economics, University of Edinburg.
Abstract: Econometricians have begun to devote more attention to spatial
interactions when carrying out applied econometric studies. The new
command we are presenting, xsmle, fits fixed- and random-effects spatial
models for balanced panel data for a wide range of specifications: the
spatial autoregressive model, spatial error model, spatial Durbin model,
spatial autoregressive model with autoregressive disturbances, and
generalized spatial random effect model with or without a dynamic
component. Different weighting matrices may be specified for different
components of the models and both Stata matrices and spmat objects are
allowed. Furthermore, xsmle calculates direct, indirect, and total
effects according to Lesage (2008), implements Lee and Yu (2010) data
transformation for fixed-effects models, and may be used with mi prefix
when the panel is unbalanced.
15:45–16:15 Estimating the dose-response function through the GLM approach
Barbara Guardabascio, Marco Ventura Italian Nationale Institute of
Statistics, Rome
Abstract: How effective are policy programs with continuous treatment
exposure? Answering this question essentially amounts to estimating a
dose-response function as proposed in Hirano and Imbens (2004). Whenever
doses are not randomly assigned but are given under experimental
conditions, estimation of a dose-response function is possible using the
Generalized Propensity Score (GPS). Since its formulation, the GPS has
been repeatedly used in observational studies, and ad hoc programs have
been provided for Stata users (doseresponse and gpscore, Bia and Mattei
2008). However, many applied works remark that the treatment variable
may not be normally distributed. In this case, the Stata programs are
not usable because they do not allow for different distribution
assumptions other than the normal density. In this paper, we overcome
this problem. Building on Bia and Mattei’s (2008) programs, we provide
doseresponse2 and gpscore, which allow one to accommodate different
distribution functions of the treatment variable. This task is
accomplished through by the application of the generalized linear models
estimator in the first step instead of the application of maximum
likelihood. In such a way, the user can have a very versatile tool
capable of handling many practical situations. It is worth highlighting
that our programs, among the many alternatives, take into account the
possibility to consistently use the GPS estimator when the treatment
variable is fractional, the flogit case by Papke and Wooldridge (1998),
a case of particular interest for economists. Predictive Margins and
Marginal Effects in Stata
16:15–16:45 Predictive Margins and Marginal Effects in Stata
Ben Jann, University of Bern [email protected]
Abstract: Tables of estimated regression coefficients, usually accompanied
by additional information such as standard errors, t-statistics,
p-values, confidence intervals or significance stars, have long been the
preferred way of communicating results from statistical models. In
recent years, however, the limits of this form of exposition have been
increasingly recognized. For example, interpretation of regression
tables can be very challenging in the presence of complications such as
interaction effects, categorical variables, or nonlinear functional
forms. Furthermore, while these issues might still be manageable in the
case of linear regression, interpretational difficulties can be
overwhelming in nonlinear models such as logistic regression. To
facilitate sensible interpretation of such models it is often necessary
to compute additional results such as marginal effects, predictive
margins, or contrasts. Moreover, smart graphical displays of results can
be very valuable in making complex relations accessible. A number of
helpful commands geared at supporting these tasks have been recently
introduced in Stata, making elaborate interpretation and communication
of regression results possible without much extra effort. Examples of
such commands are margins, contrasts, and marginsplot. In my talk, I
will discuss the capabilities of these commands and present a range of
examples illustrating their use.
16:45–17:00 Coffee
17:00–17:45 Report to the Users
Bill Rising
Abstract: Bill Rising, Director of Educational Services, talk about
developments at Stata.
17:45–18:30 Wishes & Grumbles
Workshop
========
Advanced Do-File programming and introduction to Ado-file programs
Ulrich Kohler
Description
-----------
The workshop covers two types of Stata programming: do-file programming
and adofile programming. It starts by presenting programming tools such
as local macros, extended macro functions, inline macro expansion,
loops, and branches and gives examples of their use in everyday data
analysis. The second part of the course explains the nuts and bolds of
defining a “program” in Stata. Finally the course provides a step-by-step
tutorial for implementing a user-written command into Stata
(“Ado-File”). Participants should be already familiar Stata. They should
know how to start a do-file from the command-line. Morover, command of
the data manipulation tools generate and replace, basic data-analysis
tools such as tabulate, summarize, and regress, and some experiences
with graph are necessary prerequisites. 1. Advanced do-file programming
Local macros Extended macro functions Scalars Loops 2. Ado-file
programming Program definition Parsing user input to programs
Step-by-step example Programming style
Date and Place
--------------
Thursday, June 6 2013, 9:00 – 17:00
University of Potsdam Building Nr. 6, Room S12
Campus Griebnitzsee
August-Bebel-Str. 89
14482 Potsdam
http://www.uni-potsdam.de/wiso_dekanat/
Presenter
---------
Prof. Kohler holds the chair for Methods of Empirical Social Research at
the University of Potsdam. He is co-author of "Data Analysis Using
Stata" and author of several user written Stata commands.
Fees
----
65 Euro (Workshop and Conference: 85 Euro)
Register
--------
You can enroll by contacting Anke Mrosek ([email protected]) by email
or by writing, phoning, or faxing to
Anke Mrosek
Dittrich & Partner Consulting GmbH
Prinzenstrasse 2
42697 Solingen
Germany
Tel: +49 (0)212 260 6624
Fax: +49 (0)212 260 6666
Description
-----------
The workshop covers two types of Stata programming: do-file programming
and adofile programming. It starts by presenting programming tools such
as local macros, extended macro functions, inline macro expansion,
loops, and branches and gives examples of their use in everyday data
analysis. The second part of the course explains the nuts and bolds of
defining a "program" in Stata. Finally the course provides a step-by-step
tutorial for implementing a user-written command into Stata
("Ado-File"). Participants should be already familiar Stata. They should
know how to start a do-file from the command-line. Moreover, command of
the data manipulation tools generate and replace, basic data-analysis
tools such as tabulate, summarize, and regress, and some experiences
with graph are necessary prerequisites.
1. Advanced do-file programming
- Local macros
- Extended macro functions
- Scalars
- Loops
2. Ado-file programming
- Program definition
- Parsing user input to programs
- Step-by-step example
- Programming style
*
* For searches and help try:
* http://www.stata.com/help.cgi?search
* http://www.stata.com/support/faqs/resources/statalist-faq/
* http://www.ats.ucla.edu/stat/stata/