% IMPORTANT: The following is UTF-8 encoded.  This means that in the presence
% of non-ASCII characters, it will not work with BibTeX 0.99 or older.
% Instead, you should use an up-to-date BibTeX implementation like “bibtex8” or
% “biber”.

@INPROCEEDINGS{Rathkopf:1005477,
      author       = {Rathkopf, Charles},
      title        = {{D}o {L}arge {L}anguage {M}odels {U}nderstand {M}eaning?},
      reportid     = {FZJ-2023-01493},
      year         = {2023},
      abstract     = {It is curiously difficult to articulate the capacities of
                      large language modelswithout getting yourself into
                      philosophically controversial terrain. In this talk Iexplain
                      why. The talk has three parts. In the first, I give a sketch
                      of how largelanguage models are built, with particular
                      attention to the way words arerepresented as vector
                      quantities. In the second, I describe the various ways
                      inwhich the capacities of language models have been tested
                      empirically. In thethird, I provide the main philosophical
                      argument. I argue that, in order tounderstand what large
                      language models are, we must reject the seeminglyinnocent
                      metaphysical principle that everything in the world either
                      has a mindor it does not.},
      month         = {Jan},
      date          = {2023-01-13},
      organization  = {Kimball Union Academy, online event
                       (USA), 13 Jan 2023},
      subtyp        = {Other},
      cin          = {INM-7},
      cid          = {I:(DE-Juel1)INM-7-20090406},
      pnm          = {5255 - Neuroethics and Ethics of Information (POF4-525)},
      pid          = {G:(DE-HGF)POF4-5255},
      typ          = {PUB:(DE-HGF)31},
      url          = {https://juser.fz-juelich.de/record/1005477},
}