Blog

Turning glosses into a mini corpus

Glossing Glossing is part and parcel in linguistics. Part of our training is learning to represent linguistic data in a format that can be understood by all. An example can

Defense against the Dark Art

This is part two in my series that comprises the last 1.5 years of my candidature in linguistics at NTU. The first part, #dissertating, can be found here. This installment

Hashtag Dissertating

As I’m about to do the final revisions on my dissertation, I have started reminiscing about my PhD trajectory. I envision myself writing a trilogy (or tetralogy) of posts that

The Abralin lectures

Love Linguistics in the time of cholera coronavirus. In a time where most institutions of most countries (yay Taiwan!) have closed during the Spring semester, the Brazilian Association of Linguistics,

Data packages for current and future me

tl; dr I show why it is worthwile to put my Chinese-related datasets in packages and how I went about it. Introduction I don’t know if I’m very late to

Rbootcamp 2019

tl; dr Below you find what we did during the Rbootcamp for Lexical Semanticists. In between this paragraph and the contents, there is a bit of my own #Rstory. Warning,

Tidy collostructions

tl ; dr In this post I look at the family of collexeme analysis methods originated by Gries and Stefanowitsch. Since they use a lot of Base R, and love

Guanguan goes the Chinese Word Segmentation (II)

tl; dr This double blog is first about the opening line of the Book of Odes, and later about how to deal with Chinese word segmentation, and my current implementation

Guanguan goes the Chinese Word Segmentation (I)

tl; dr This double blog is first about the opening line of the Book of Odes, and later about how to deal with Chinese word segmentation, and my current implementation

Memorable moments of 2018

Hi, reader. Since I am doing research most of the time and somewhat struggling with the question of what I should let out of the bag and on this blog,