Follow HLP lab on the English Zwitscher
Only a few years (decades?) late, HLP lab is now zwitschering insanely uninteresting things on Twitter. You can follow us and get updates about workshops, classes, papers, code, etc. And you can zwitscher back at us and we can all be merry and follow and comment on each other until our eyes pop out or ears explode. In this spirit: @_hlplab_
Presentation at CNS symposium on “Prediction, adaptation and plasticity of language processing in the adult brain”
Earlier this week, Dave Kleinschmidt and I gave a presentation as part of a mini-symposium at Cognitive Neuroscience Conference on “Prediction, adaptation and plasticity of language processing in the adult brain” organized by Gina Kuperberg. For this symposium we were tasked to address the following questions:
- What is prediction and why do we predict?
- What is adaptation and why do we adapt?
- How do prediction and adaptation relate?
Although we address these questions in the context of language processing, most of our points are pretty general. We aim to provide intuitions about the notions of distribution, prediction, distributional/statistical learning and adaptation. We walked through examples of belief-updating, intentionally keeping our presentation math-free. Perhaps some of the slides are of interest to some of you, so I attached them below. A more in-depth treatment of these questions is also provided in Kleinschmidt & Jaeger (under review, available on request).
Comments welcome. (sorry – some of the slides look strange after importing them and all the animations got lost but I think they are all readable).
It was great to see these notions discussed and related to ERP, MEG, and fMRI research in the three other presentations of the symposium by Matt Davis, Kara Federmeier and Eddy Wlotko, and Gina Kuperberg. You can read their abstracts following the link to the symposium I included above.
Post-doctoral position available (speech perception, language comprehension, implicit distributional learning, inference under uncertainty, hierarchical predictive systems)
The Human Language Processing (HLP) Lab at the University of Rochester is looking for a post-doctoral researcher interested in speech perception and adaptation. Possible start dates for this 1-3 year position range from mid August 2014 to mid June 2015 (the current post-doctoral researcher funded under this grant will leave HLP lab in late August to start a tenure-track position in Psychology at the University of Pittsburgh). International students are welcome to apply (NIH research grants are not limited to nationals).
We will start reviewing applications mid-June 2014 though later submissions are welcome. Applications should contain (1) a cover letter clearly indicated possible start dates, (2) a CV, (3) research statement detailing qualifications and research interests, and (4) 2 or more letters of recommendation. Applications and letters should be emailed to Kathy Corser (kcorser@bcs.rochester.edu), subject line “application for post-doc position (HLP Lab)”.
This is an NIH funded project (NIHCD R01 HD075797), currently scheduled to end in 2018. The project is a collaboration between Florian Jaeger (PI), Mike Tanenhaus (co-PI), Robbie Jacobs and Dick Aslin. We are interested in Read the rest of this entry »
A good start to Klinton Bicknell at Northwestern
A belated congratulations and a good start to Klinton Bicknell, who just started his tenure-track position in Linguistics at Northwestern. To quote his website, his research:
… seeks to understand the remarkable efficiency of language comprehension, using the tools of probability theory and statistical decision theory as explanatory frameworks. My work suggests that we achieve communicative efficiency by utilizing rich, structured probabilistic information about language: leveraging linguistic redundancy to fill in details absent from the perceptual signal, to spend less time processing more frequent material, and to make predictions about language material not yet encountered.
old and new lme4
(This is a guest post by Klinton Bicknell.)
update 2014-06-24: Using lme4.0 probably isn’t necessary anymore. See post here.
The lme4 package‘s major 1.0 release was back in August. I and others have noticed that for typical psycholinguistic datasets, the new >=1.0 versions of lme4 often yield models with substantially poorer fits to the data than the old pre-1.0 versions (sometimes worse by many points of log likelihood), which suggests that the new lme4 isn’t as reliably converging to the actual maximum likelihood (or REML) solution. Since unconverged models yield misleading inferences about model parameters, it’s useful to be able to fit models using the old pre-1.0 lme4.
Happily, the lme4 developers have created a new package (named “lme4.0”), which is a bugfix-only version of the old pre-1.0 lme4. This allows for the installation of both old and new versions of lme4 side-by-side. As of this posting, lme4.0 is not yet on CRAN, but is installable by performing the following steps: Read the rest of this entry »
Another example of recording spoken productions over the web
A few days ago, I posted a summary of some recent work on syntactic alignment with Kodi Weatherholtz and Kathryn Campell-Kibler (both at The Ohio State University), in which we used the WAMI interface to collect speech data for research on language production over Amazon’s Mechanical Turk.
Socially-mediated syntactic alignment
The first step in our OSU-Rochester collaboration on socially-mediated syntactic alignment has been submitted a couple of weeks ago. Kodi Weatherholtz in Linguistics at The Ohio State University took the lead in this project together with Kathryn Campbell-Kibler (same department) and me.
We collected spoken picture descriptions via Amazon’s crowdsourcing platform Mechanical Turk to investigate how social attitude towards an interlocutor and conflict management styles affected syntactic priming. Our paradigm combines Read the rest of this entry »
Ways of plotting map data in R (and python)
Thanks to Scott Jackson, Daniel Ezra Johnson, David Morris, Michael Shvartzman, and Nathanial Smith for the recommendations and pointers to the packages mentioned below.
- R:
- The maps, mapsextra, and maptools packages provide data and tools to plot world, US, and a variety of regional maps (see also mapproj and mapdata). This, combined with ggplot2 is also what we used in Jaeger et al., (2011, 2012) to plot distributions over world maps. Here’s an example from ggplot2 with maps.
Join me at the 15th Texas Linguistic Society conference?
I’ll be giving a plenary presentation at the 15th Texas Linguistic Society conference to be held in October in Austin, TX. Philippe Schlenker from NYU and David Beaver from Austin will be giving plenaries, too. The special session will be on the “importance of experimental evidence in theories of syntax and semantics, and focus on research that highlights the unique advantages of the experimental environment, as opposed to other sources of data” (from their website). Submit an abstract before May 1st and I see you there.
Perspective paper on second (and third and …) language learning as hierarchical inference
We’ve just submitted a perspective paper on second (and third and …) language learning as hierarchical inference that I hope might be of interest to some of you (feedback welcome).
- Pajak, B., Fine, A.B., Kleinschmidt, D., and Jaeger, T.F. submitted. Learning additional languages as hierarchical probabilistic inference: insights from L1 processing. submitted for review to Language Learning.
We’re building on Read the rest of this entry »
Your thoughts on second, third, etc. language learning
After a recent discussion in the lab, I’m curious to hear your speculative or informed opinions about the factors that determine someone’s proficiency in a second (or third, …) language. I’ve put together a brief survey (5-10 minutes) and if you have some time, it’d be great to hear your thoughts. Experts’ and non-experts’ opinions are equally welcome. I will post the results of the survey here.
Thank you for your participation.
And a belated welcome to Scott Grimm
Scott Grimm just joined our faculty in Linguistics at Rochester last month. So today he got his belated Rochester welcome:
Scott joins the Center of Language Sciences at Rochester with an unnecessary number of degrees. Read the rest of this entry »
Updated slides on GLM, GLMM, plyr, etc. available
Some of you asked for the slides to the Mixed effect regression class I taught at the 2013 LSA Summer Institute in Ann Arbor, MI. The class covered some Generalized Linear Model, Generalized Linear Mixed Models, extensions beyond the linear model, simulation-based approaches to assessing the validity (or power) of your analysis, data summarization and visualization, and reporting of results. The class included slides from Maureen Gillespie, Dave Kleinschmidt, and Judith Degen (see above link). Dave even came by to Ann Arbor and gave his lecture on the awesome power of plyr (and reshape etc.), which I recommend. You might also just browse through them to get an idea of some new libraries (such as Stargazer for quick and nice looking latex tables). There’s also a small example to work through for time series analysis (for beginners).
Almost all slides were created in knitr and latex (very conveniently integrated into RStudio — I know some purists hate it, but comm’on), so that the code on the slides is the code that generated the output on the slides. Feedback welcome.
Workshop announcement (Tuebingen): Advances in Visual Methods for Linguistics
This workshop on data visualization might be of interest to a lot of you. I wish I could just hop over the pond.
- Date: 24-Sept-2014 – 26-Sept-2014
- Location: Tuebingen, Germany
- Contact Person: Fabian Tomaschek (contact@avml-meeting.com)
- Web Site: http://avml-meeting.com
- Call Deadlines: 21 March / 18 April
The AVML-meeting offers a meeting place for all linguists from all fields who are interested in elaborating and improving their data visualization skills and methods. The meeting consists of a one-day hands-on workshop Read the rest of this entry »
Is my analysis problematic? A simulation-based example
This post is in reply to a recent question on in ling-R-lang by Meredith Tamminga. Meredith was wondering whether an analysis she had in mind for her project was circular, causing the pattern of results predicted by the hypothesis that she was interested in testing. I felt her question (described below in more detail) was an interesting example that might best be answered with some simulations. Reasoning through an analysis can, of course, help a lot in understanding (or better, as in Meredith’s case, anticipating) problems with the interpretation of the results. Not all too infrequently, however, I find that intuition fails or isn’t sufficiently conclusive. In those cases, simulations can be a powerful tool in understanding your analysis. So, I decided to give it a go and use this as an example of how one might approach this type of question.
- ← Previous
- 1
- 2
- 3
- …
- 14
- Next →