{
  "_schema": "https://data.nist.gov/od/dm/nerdm-schema/v0.7#",
  "@context": [
    "https://data.nist.gov/od/dm/nerdm-pub-context.jsonld",
    {
      "@base": "ark:/88434/mds2-3074"
    }
  ],
  "@type": [
    "nrdp:PublicDataResource",
    "dcat:Dataset"
  ],
  "_extensionSchemas": [
    "https://data.nist.gov/od/dm/nerdm-schema/pub/v0.7#/definitions/PublicDataResource"
  ],
  "@id": "ark:/88434/mds2-3074",
  "ediid": "ark:/88434/mds2-3074",
  "version": "1.0.0",
  "doi": "doi:10.18434/mds2-3074",
  "title": "Meta Learning Paper Supplemental Code",
  "contactPoint": {
    "fn": "Apostol Vassilev",
    "hasEmail": "mailto:apostol.vassilev@nist.gov"
  },
  "modified": "2023-09-11",
  "status": "available",
  "landingPage": "https://data.nist.gov/od/id/mds2-3074",
  "description": [
    "Meta learning with LLM: supplemental code for reproducibility of computational results for MLT and MLT-plus-TM. Related research paper: \"META LEARNING WITH LANGUAGE MODELS: CHALLENGES AND OPPORTUNITIES IN THE CLASSIFICATION OF IMBALANCED TEXT\", A. Vassilev, H. Jin, M. Hasan, 2023 (to appear on arXiv).",
    "All code and data is contained in the zip archive arxiv2023.zip, subject to the licensing terms shown below. See the Readme.txt contained there for detailed explanation how to unpack and run the code. See also requirements.txt for the necessary depedencies (libraries needed). This is not a dataset, but only python source code."
  ],
  "keyword": [
    "Natural language processing",
    "Out of policy speech detection",
    "Meta learning",
    "Deep learning",
    "Language Models"
  ],
  "topic": [
    {
      "@type": "Concept",
      "scheme": "https://data.nist.gov/od/dm/nist-themes/v1.1",
      "tag": "Mathematics and Statistics"
    },
    {
      "@type": "Concept",
      "scheme": "https://data.nist.gov/od/dm/nist-themes/v1.1",
      "tag": "Information Technology: Computational science"
    }
  ],
  "accessLevel": "public",
  "license": "https://www.nist.gov/open/license",
  "publisher": {
    "name": "National Institute of Standards and Technology",
    "@type": "org:Organization"
  },
  "language": [
    "en"
  ],
  "bureauCode": [
    "006:55"
  ],
  "programCode": [
    "006:052"
  ],
  "theme": [
    "Mathematics and Statistics",
    "Information Technology: Computational science"
  ],
  "rights": "N/A",
  "components": [
    {
      "accessURL": "https://github.com/usnistgov/NIST-AI-Meta-Learning-LLM",
      "format": {
        "description": "Zip archive"
      },
      "description": "Meta learning with LLM: supplemental code for reproducibility of computational results for MLT and MLT-plus-TM. Related research paper: \"META LEARNING WITH LANGUAGE MODELS: CHALLENGES AND OPPORTUNITIES IN THE CLASSIFICATION OF IMBALANCED TEXT\", A. Vassilev, H. Jin, M. Hasan, 2023 (to appear on arXiv).\n\nAll code and data is contained in the zip archive arxiv2023.zip, subject to the licensing terms shown below. See the Readme.txt contained there for detailed explanation how to unpack and run the code. See also requirements.txt for the necessary depedencies (libraries needed).",
      "title": "Meta Learning Paper supplemental code",
      "@type": [
        "nrdp:AccessPage",
        "dcat:Distribution"
      ],
      "@id": "#usnistgov/NIST-AI-Meta-Learning-LLM",
      "_extensionSchemas": [
        "https://data.nist.gov/od/dm/nerdm-schema/pub/v0.7#/definitions/AccessPage"
      ]
    },
    {
      "@id": "cmps/arxiv2023.zip",
      "@type": [
        "nrdp:DataFile",
        "nrdp:DownloadableFile",
        "dcat:Distribution"
      ],
      "_extensionSchemas": [
        "https://data.nist.gov/od/dm/nerdm-schema/pub/v0.7#/definitions/DataFile"
      ],
      "filepath": "arxiv2023.zip",
      "downloadURL": "https://data.nist.gov/od/ds/mds2-3074/arxiv2023.zip",
      "mediaType": "application/zip",
      "title": "Meta Learning Paper supplemental code",
      "size": 4218084,
      "checksum": {
        "hash": "a5a774fe0304da834d9be825a34696c5fe5d19fe47f86b24c4753f83a81cd446",
        "algorithm": {
          "tag": "sha256",
          "@type": "Thing"
        }
      },
      "format": {
        "description": "zip archive of text files (pyhhon source code)"
      },
      "description": "Meta learning with LLM: supplemental code for reproducibility of computational results for MLT and MLT-plus-TM. Related research paper: \"META LEARNING WITH LANGUAGE MODELS: CHALLENGES AND OPPORTUNITIES IN THE CLASSIFICATION OF IMBALANCED TEXT\", A. Vassilev, H. Jin, M. Hasan, 2023 (to appear on arXiv).\n\nAll code and data is contained in the zip archive arxiv2023.zip, subject to the licensing terms shown below. See the Readme.txt contained there for detailed explanation how to unpack and run the code. See also requirements.txt for the necessary depedencies (libraries needed)."
    }
  ],
  "annotated": "2023-10-13T15:45:54.443244",
  "revised": "2023-10-13T15:45:54.443244",
  "issued": null,
  "firstIssued": "2023-10-13T15:45:54.443244"
}