Thursday, February 20, 2025

In the boat together

There's a prolific scene in the movie Titanic where the band continues to play while the ship is sinking. If they know the ship is sinking, do they continue because their employers want it? because they are avoiding such terror by distracting themselves with something banal? because of their commitment to making music in spite of their world ending? because it is all they could choose to do in the face of terror?

I'm caught in how much this is an apt scene for how things feel right now. Is the boat sinking for research and scholarship in the United States? If so, what motivates the choices we make to continue doing what we're trained to do? what we're passionate about? Do these things matter relative to the fear of the moment?

Universities would have you believe that what remains motivating is some sort of commitment to the institution or "because you're paid to do it." That's bs, of course. Certain institutions might engender such faith, but most don't. Are we free to choose our passions at this moment then? or does that seem selfish relative to all the other areas where we can choose to protest or serve people?
(I realize in writing this that the notion of doing something for yourself being equated with selfishness is a persistent affliction of me growing up Catholic that I still have to kick.)
Universities would also have you believe that you find motivation in the "big picture" questions in your field. These things can be inspiring in the short term, but are often not that motivating in the long run. Instead, it seems to be how much we connect with others around our interests that seems to motivate me. Am I regularly making progress on some interesting research topic? Am I spending time each week talking about my field with colleagues? with other experts? with long-time friends? Am I helping students in a lab work on their data analysis issues? These are also the things providing me some sense of purpose.

And so maybe it isn't a sense of duty that kept the band playing on the bridge of the Titanic, or some vague notion that they were committed to beauty amidst terror. Maybe it was just a commitment to support each other still playing that gave them motivation and purpose. Maybe it is always about others.

Tuesday, December 24, 2024

A retention of faith

What does it mean to retain one's faith? Is it faith in oneself? in your community? in society? in others? in a deity of some sort? Do we have to have faith and trust in ourselves before we have it in our communities? Or is it the other way around? Might we not have faith in ourselves if we see no examples of it around us?

Following the election, I am troubled by what is a gradual unfolding of that faith in my country, in institutions that are supposed to be motivated by justice and common good instead of by avarice and personal vengeance. If this country is only about greed and selfishness, then I gradually find myself at odds with it, a stranger floating adrift among other strangers. And if I find community among strangers, how do you re-form (or reform) a community on the basis of beliefs you all thought everyone shared?

Does that crisis of faith in some abstract type of thing - a country as a concept - trickle down into other abstract notions like society? Does it trickle down into things of substance like communities or the people at the grocery store? to my students? to my family?

There are the communities I live in - in New Haven, in Buffalo, among faculty and students at UB, among friend groups in different parts of the world, among family. I have needed to touch each community to check again and again if they remain intact, clinging onto the fabric for a bit asking "Do you also still believe in kindness? in noble pursuits?"

And if selfishness abounds while knowledge languishes at the margins, I find myself struggling against so much lost hope. Do I believe that students are fundamentally interested in learning still? Do I believe that I can do anything to prevent academic dishonesty in a world of easy AI answers? Should I even bother trying? Do I believe that work in my research and my field is worth pursuing while universities, institutions, and society (that diaphanous term) might flippantly dismiss such endeavors as wasteful or impractical? Why does it feel like my careful little efforts are like planting trees in a forest fire?

Or does the despair and lack of faith only trickle up? Does the individual losing faith become nothing but a black hole sucking all others into their vortex? Is this the nature of where faith goes? And as communities become lost to these cults of individualism, what is left to hold onto?

Monday, December 2, 2024

Research

Did you discuss your project proposal? Have you talked about it with colleagues? Have you revised it? Have you mulled it over in the shower, on your walk, in the wee hours of the morning?

Have you requested funding from a grant source? Did it take you 1-2 months to write a proposal and work with your university to submit all the paperwork in time? Did you wait 6 months to hear back about whether you got funded? If you did get funded, did you wait another 3 months to start the research?

Did you write up your protocol for the institutional review board at your university (if the work involves human subjects)? Did you have to revise this? Did it take you 3-4 weeks and did you pull your hair out trying to answer their questions?

Once you had the research funds, did you train students as research assistants? did you design experiments and run them? Have you finished collecting all your data? Did you stay late at the university again doing so?

Did you spend 6-9 months analyzing and re-analyzing your data, plotting your findings, doing statistics? Did you present these new findings at research conferences that you submitted to and then attended? Did your university not give you enough funds to attend the conference? Have you incorporated feedback from your colleagues now that you have findings?

Have you written up your results in a research paper and submit it to a peer-reviewed journal? Have you read the 10, 20, 30, 40 relevant articles that you cite in your background section? Are your findings well argued?

Did you wait 3-4 months to hear back from the journal? If the journal's decision was "revise and resubmit" or "major revisions", did you spend 2-3 months revising your manuscript to address the reviewers' concerns? Did you pull your hair out doing this too?

Did you repeat this process above to revise it yet again? Did the journal give you reviewers that advised you to do more than last time? Do you have any hair left?

If your paper was accepted after the 9-12 months for journal submission, review, and revision, did you also address all copyeditor commentary?

If this all has happened, congratulations, you have published an original piece of research. It took you 2-3 years and a huge amount of effort. Hopefully you can continue to do this on your grant.

Now, remind me, what was that again about your google "research" that you spent 30 minutes on?

Wednesday, November 20, 2024

November thoughts

11/19/24

I believed that I could put aside my anxious spiraling, catastrophizing thoughts into a box. I mostly succeeded in doing so. I quit social media a week before the election. It's now been about 3.5 weeks without Facebook. I eliminated my Twitter account. I've eliminated NY Times (about a month ago). I don't read the news. I actively try to avoid it.

I've embraced the new life that has come into Bsky (blue sky), but it's with some trepidation. If people post about politics, either via worry about the incoming president's staff picks or worrying about potential laws and actions, I can't help thinking about it throughout the day. It's certainly not as paralyzing as my dread in October was, but it's creeping in that direction.

I find that I'm still left with these strong feelings about politics. I get angry at the idea of completely unprofessional people being in charge of things. I get angry at how much I fear I will have to fight to prove my value, or to prove that my values are worth preserving. I worry I might have to spend years of my time in the future fighting a system that will eventually collapse anyways, all while my gentle, curious mind is snuffed out by all the chaos and noise. Will I grow to believe that research is futile? Is there no place for introverts in the future of America?

All these things have been on my mind of late. They're theoretical scenarios, I know, so nothing is real yet. But believing in what is happening, I find that I am forced to make some very negative conclusions about Americans - ones that I don't feel that I made in the past. This is what concerns me.

I used to feel that people were simply ill-informed about a lot of issues. That probably has always been true, but I find myself believing that people don't care about what is true nowadays. That's more concerning. And where my brain goes with this is to believe that others don't care about knowledge, just power. If knowledge has no access to power, then in a selfish world, it never makes sense to pursue it. You're left concluding that knowledgeable people are expendable. I'm also left concluding that most people believe I don't matter.

I feel like I used to believe that, fundamentally, people should care for each other. Show someone examples of how others struggle or talk to people about your life and they will empathize. It's human nature. Yet, we're given a political landscape where people have excuses to reject most other people. There are excuses to dislike immigrants. There are excuses to dislike Latinos and Black people. There are excuses to dislike LGBTQ folks. Accept all of them and you're left with people who have drunk the kool-aid of structural prejudice and dog-whistle politics. And it's not that I didn't believe that prejudiced people existed before, it's that I believed that they would want to change. If I now believe people don't want to change or learn about others, then I have trouble believing in their fundamental goodness. 

Taken to its conclusion, if no one wants to change their views about the world or others, then they learn nothing from me. After all, it makes no sense to engage with people who have no interest in learning from others. Dark thoughts, surely.

So, without social media, as I'm not anxiously spiraling worrying about the future, my brain has still made what seems to me to be very logical conclusions about the nature of people. I suppose that the scientist in me wants proof that most people are still fundamentally good, while the child in me wants to believe that I don't live in a country as hateful as I fear it is.

Sunday, December 10, 2023

On the generalization of linguistic discovery

Discovery is a crucial part of the evolution of most academic disciplines that take a scientific approach towards understanding the world. New empirical evidence of a phenomenon leads researchers to re-examine old perceptions they had. Or rather, as Kuhn (1962) would argue, those with the old perceptions of the world eventually die or fade away while those who only have these newer perceptions mature.

But how do we generalize discovery? There are certainly many disciplines where discovery is generalizable. Findings in many of the physical sciences and mathematics are truths that will continue to be true forever. Discover a solution to a long-held mathemetical problem and it will remain true from now on. 

In the social and cognitive sciences though, discoveries seem somewhat murkier. Where they relate to biological, neurobiological, biophysical principles, the discoveries seem more generalizable. In my main sub-discipline, phonetics, there are clear physical relationships between what a person does with their speech articulators and what this produces in an acoustic signal, for instance. This is true across languages because all humans have similar oral and laryngeal anatomy. Yet, since speakers can massively vary just how they produce similar speech sounds, generalization is challenging here too.

Where they do not relate to biological or physical principles, behavioral and linguistic discoveries are usually observational findings restricted to a certain type of population. Generalization here necessarily needs to proceed to multiple experiments or studies with different types of populations. From a linguists' perspective (and I can only speak as a linguist here), that necessarily means that discoveries need more languages. 

There's a danger here that comes out of a kind of science-envy with behavioral and linguistic sciences. Though some of the methods in the social/behavioral sciences have become more scientifically rigorous (mostly in relation to statistical testing and modeling), the findings are not magically more generalizable to new populations than they were in the past. Discovering that college-aged speakers of English prefer certain syntactic structures over others does not mean anything about any other language unless subsequent research is undertaken. It might make predictions about patterns in other languages, but predictions are not generalizations.

Can we ever generalize about "Language"? What if we can't?

There are a lot of half-truths that linguists hold about "Language" that arise from a casual extension of findings in a few languages. Demonstrate that some linguistic phenomenon occurs in American English, Spanish, and German and linguists will believe it is a universal or "strong tendency" without a very clear criterion for what "universal" or "strong tendency" would mean.

Why be so careful with formal and statistical methods but so uncareful regarding the scientific bread-and-butter of hypothesis testing? The answer seems to lie in a kind of all-or-nothing perspective about where linguistic discoveries have value to a discipline. Linguists either believe linguistic patterns demonstrate unique characteristics of individual languages or populations; -or- they are universal patterns reflecting something deep about human evolution or murkier things like universal grammar. The field tends to narrowly merit the latter type of work since it is smells like a generalization.

This all-or-nothing approach means that we often come up empty-handed when we wish to talk about the relevance of our findings to the discipline - we're delving deeply into specific languages with an empirical or historical goal or we're looking broadly (and more superficially) at patterns in a larger number of languages. What might exist in the middle? We're a small discipline examining a huge topic with a gigantic amount of variation. We can't do it all.

I think one future path for the discipline is to take a note from the quantitative revolution that has occurred over the past 20-25 years in the discipline. The more we examine phenomena that we once believed to be discrete (x occurs in context A, but y occurs in context B), the more we discover that these are strong statistical tendencies instead. And the reason for this is that linguistic phenomena are behavioral. They are not the formal mathematical proofs that remain true forever after being solved. We just keep wanting to commit our error of generalization because of this science envy.

Might there not be any true linguistic universals? Maybe there are but we can never be typologically-balanced enough to prove anything more than fairly superficial patterns. Maybe there aren't any at all and this is ok. Languages are endlessly fascinating and we can still demonstrate how many languages work along statistical lines. The idea that there is massive inter-language variation and that this is structured to occur in certain types of languages necessarily means that we can look at types of languages to construct complex cross-linguistic hypotheses. To provide a concrete example, do speakers of fusional languages or those with non-concatentive morphology store words differently than speakers of isolating languages? This is an interesting question but it does not require a model of what must be universal. It just requires experiments and cross-linguistic research.

This is a blog post, so take my musings with a grain of salt. I don't have the answers to my own subdiscipline, let alone all of linguistics. I think though that we need to be more careful distinguishing between the things that we believe are proven/demonstrated and the things that are demonstrated typological patterns or universals. 

Wednesday, November 1, 2023

Issues in choosing a statistical model in phonetics

What's the bar for deciding to use a new statistical model in research? It seems like often enough within linguistics or speech science, one chooses a model based on what is à la mode. That frequently translates into increasing complexity.

Is it always good to have a more complex model? No. It might reveal more intricate interactions in the data. It might also model interactions between terms better than competing models, usually by improving fit with non-linear terms (cf. GCA, GAMMS). Yet, there are missing evaluative criteria for choosing a model that end up being crucially important. 

1. Is the model easily implementable and understandable? 

If a model is easy to implement and understand, then it is easy enough for new users to emerge and for a set of standards to come about. Yet, if neither of these things are true, there is severe social cost. 

If there are a handful of researchers proposing using a new model, is there an existing infrastructure that can help with training and implementation? Usually there is not and, as a consequence, many researchers get frustrated if the field pushes a model where no infrastructure exists. The same people proposing the model will end up fielding hundreds or thousands of questions about how to use it. And nobody has time for that.

Now, why might the field (or paper reviewers, most likely) decide that everyone has to use one particularly new and popular model for one's data? Sometimes important new factors are discovered that need to be modeled. But sometimes it's just the impostor syndrome, i.e. we are only a serious field if we have increasingly more mathematically opaque models for our data. And it's easy to give a post-hoc reason to include all possible factors when our predictions are so weak.

2. Does the model enable us to generalize?

Do we actually need to model as many of the details as we can? Even models that take a fairly generic approach to avoiding overfitting can end up overfitting things like dynamics. Resultingly, researchers lose time needing to discuss details that end up being unimportant and we end up losing the ability to generalize.

I'll provide one personal example of this. In my co-authored paper on the phonetics of focus in Yoloxóchitl Mixtec, we provided statistical models for f0 dynamics alongside statistical models for midpoint f0 values. There is certainly good reason to model changes in f0, but in a language with a number of level tones (and tone levels), this type of modeling might not say much. Indeed, we found mostly the same results when we looked at f0 midpoint for many of the level tones than when we looked at dynamic trajectories for them. Including two sets of models resulted in twice as many statistical tests and twice as much reporting.

Why did we choose to do this? We favored being comprehensive over possibly missing some unknown pattern (maybe the lower level tones had some different dynamic behavior?) Given the subtlety of the resulting patterns, it's hard to say what might be important.

Nowadays, I think we would be asked to choose to use GAMS instead of the mixed effects modeling. Yet, that also results in a statistical bloat (e.g. you have to model each tone separately). The results of our research should lead us to make scientific conclusions about speech, not get lost in 101 statistical tests where we spend time analyzing our three-way interactions. 

I don't know the right answer to how the field might address this issue, but I do not believe that it has to do with reducing the purview of one's study. GAMs are great if you are looking at one pattern in one language, but they are terrible for generalizing over a language's inventory (of vowel formants, of tones, of prosodic contexts, etc). One finds either studies using GAMs for limited topics (one vowel or one context) or studies where 101 models are included to provide a comprehensive account of a language's patterns. The former are more likely in studies examining well-studied languages while the latter are more likely in exploratory analyses of languages.

The negative consequence here might be that the "clear case" for GAMs is made within the less complex pattern in a well-studied language, while no one can make heads or tails of all the analyses in the less well-studied language. I see this as just an extension of linguistic common ground as privilege. Yet, now it's done with statistics.




Sunday, January 8, 2023

The "Bender rule" in some linguistics journals in 2022

The Bender rule is the informal idea that one ought to explicitly mention the name of a language in a publication on language and linguistics. It is named after Emily Bender, a computational linguist at the University of Washington (Seattle) who has written and discussed the need to be explicit about languages that one studies. The impetus behind it is the observation that studies on English (or other commonly-studied languages) are typically understood as a default norm, while less commonly studied languages are more likely to be overtly mentioned. This contributes to a biased perspective in linguistics that only the conclusions from studies on English contribute to a general picture of Language, while similar conclusions from studies on other languages reflect language-specific phenomena and are less generalizable. A similar issue arises in work on indigenous languages that I've written about before.

People have talked about the Bender rule since 2019. I'd like to think that linguists have paid attention to what this means in academic publications since then. After all, it would be fairly simple for journal editors or editorial boards to implement a policy where languages are mentioned in titles or in abstracts. After all, people often read/skim the titles and abstracts of most publications without investing in more time to read all the details. If one were to apply the Bender rule to titles and/or abstracts (and yes, I am suggesting it), it has the additional benefit of helping your librarians organize publications better by topic language.

So, how have some popular journals fared in 2022? Are many publications mentioning the languages of study? I thought I would look at two popular journals that I am familiar with: the Journal of Memory and Language (JML), and the Journal of Phonetics (JPhon). Both journals heavily focus on experimental research. I decided to include two separate measures here: does the journal article mention the language of study in the title? and does it mention it in the abstract? I have excluded publications that reflect surveys of methodological reports, as these lack experiments and they tend not to focus on individual languages anyways.

For JML, between January 2022 - present, 43 relevant articles have been published. Of these, just 2/43 mention the language of study in the title. Within the abstracts, 8/43 articles mention the language of study. Studies that explicitly mentioned languages were those on Mandarin Chinese, ASL, and those involving bilingual populations.

For JPhon, between January 2022 - present, 40 relevant articles have been published. Of these, 18/40 mentioned the language of study in the title. Within the abstracts, 35/40 articles mention the language of study. 

Why might these numbers (and practices) might be so different across journals? Are the psycholinguistic patterns found in brains and minds in the articles in JML fundamentally different in terms of their language-specificity from studies on phonetic memory/perception, speech planning, speech coordination, and speech articulation found in JPhon? In other words, is it that only the phoneticians need worry about the Bender rule?

I think most phoneticians would probably state that a study on the articulatory and acoustic phonetics of one language is bound to be fundamentally different from a similar study on another language. Thus, there is less of an expectation that one's findings will immediately generalize to all of Language. Rather, one draws conclusions and amasses evidence for common patterns by looking across a large enough sample of languages. Existing theories are examined, tested with new data, and revised.

I don't know what psycholinguists believe here though. Perhaps it is the case that many still believe that English-focused studies in psycholinguistics are always uncovering something fundamental about Language in a way that studies in phonetics are not, despite apparent evidence to the contrary. I have to doubt that though. I know many psycholinguists and they seem to be a pretty open-minded group. For the time being, it would seem like JML is failing the Bender rule.