{"payload":{"pageCount":10,"repositories":[{"type":"Public","name":"vigor","owner":"amazon-science","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":0,"license":"Other","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-12T22:28:48.252Z"}},{"type":"Public","name":"learning-to-generate-answers-with-citations","owner":"amazon-science","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":1,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-12T00:15:12.749Z"}},{"type":"Public","name":"ssepy","owner":"amazon-science","isFork":false,"description":"Python package for stratifying, sampling, and estimating model performance efficiently.","allTopics":["statistics","estimation","statistical-inference","sampling","stratified-sampling"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":0,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,12],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-11T04:50:20.460Z"}},{"type":"Public","name":"collage","owner":"amazon-science","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-10T19:29:36.657Z"}},{"type":"Public","name":"RAGChecker","owner":"amazon-science","isFork":false,"description":"RAGChecker: A Fine-grained Framework For Diagnosing RAG","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":12,"forksCount":0,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,11,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-10T17:32:38.774Z"}},{"type":"Public","name":"RefChecker","owner":"amazon-science","isFork":false,"description":"RefChecker provides automatic checking pipeline and benchmark dataset for detecting fine-grained hallucinations generated by Large Language Models.","allTopics":["hallucination","factuality","llms"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":6,"starsCount":226,"forksCount":17,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,5,0,5,0,2,0,10,6,5,0,0,1,1,8,7,3,0,3,0,0,0,2,1,0,0,1,0,1,8,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-10T17:31:36.438Z"}},{"type":"Public","name":"graph-lm-ensemble","owner":"amazon-science","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":10,"forksCount":3,"license":"Apache License 2.0","participation":[2,1,0,1,1,6,1,2,0,0,0,2,3,1,1,1,1,0,0,1,1,0,1,0,0,0,13,0,0,0,1,0,0,0,0,0,0,1,2,1,0,0,0,0,0,0,0,0,0,0,0,22],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-10T17:10:33.363Z"}},{"type":"Public","name":"patchcore-inspection","owner":"amazon-science","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":71,"starsCount":680,"forksCount":142,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-10T01:35:09.912Z"}},{"type":"Public","name":"avgen-eval-toolkit","owner":"amazon-science","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":1,"forksCount":2,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-09T18:46:39.435Z"}},{"type":"Public","name":"BYOKG-NAACL24","owner":"amazon-science","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":7,"forksCount":1,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,27,3,1,0,1,0,0,1,0,1,0,1,0,0,9],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-09T18:46:19.622Z"}},{"type":"Public","name":"text_generation_diffusion_llm_topic","owner":"amazon-science","isFork":false,"description":"Topic Embedding, Text Generation and Modeling using diffusion","allTopics":["nlp","machine-learning","natural-language-processing","topic","transformers","text-generation","topic-modeling","lda","text-embedding","topic-models","sentence-embeddings","diffusion-models","t5","text-embeddings"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":7,"forksCount":1,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,5,0,1,4,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-08T22:04:52.709Z"}},{"type":"Public","name":"ccsum","owner":"amazon-science","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":3,"forksCount":0,"license":"Other","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,1,2,0,0,0,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-08T18:20:48.553Z"}},{"type":"Public","name":"cocomic","owner":"amazon-science","isFork":false,"description":"CoCoMIC: Code Completion By Jointly Modeling In-file and Cross-file Context","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":0,"starsCount":6,"forksCount":2,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-07T04:33:26.516Z"}},{"type":"Public","name":"tree-of-traversals","owner":"amazon-science","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-07T04:15:21.718Z"}},{"type":"Public","name":"synthesizrr","owner":"amazon-science","isFork":false,"description":"Synthesizing realistic and diverse text-datasets from augmented LLMs","allTopics":["deep-learning","language-modeling","natural-language-generation","synthetic-data","synthetic-dataset-generation","retrieval-augmented","retrieval-augmented-generation"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":2,"forksCount":1,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-06T02:28:27.142Z"}},{"type":"Public","name":"doc-mt-metrics","owner":"amazon-science","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":17,"forksCount":3,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T21:37:57.212Z"}},{"type":"Public","name":"mada_optimizer_search","owner":"amazon-science","isFork":false,"description":"Code the ICML 2024 paper: \"MADA: Meta-Adaptive Optimizers through hyper-gradient Descent\"","allTopics":["machine-learning","deep-neural-networks","optimization","machine-learning-algorithms","optimization-algorithms","adam-optimizer","gpt-2","meta-optimizer","large-language-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":3,"forksCount":1,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-03T20:37:49.071Z"}},{"type":"Public","name":"factual-confidence-of-llms","owner":"amazon-science","isFork":false,"description":"Code for paper \"Factual Confidence of LLMs: on Reliability and Robustness of Current Estimators\"","allTopics":["robustness","factual","confidence","factuality","llm","llms"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":0,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-03T12:42:05.177Z"}},{"type":"Public","name":"Repoformer","owner":"amazon-science","isFork":false,"description":"Repoformer: Selective Retrieval for Repository-Level Code Completion (ICML 2024)","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":3,"starsCount":26,"forksCount":3,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-01T11:07:00.097Z"}},{"type":"Public","name":"path-consistency","owner":"amazon-science","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":1,"starsCount":8,"forksCount":3,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-30T10:36:30.860Z"}},{"type":"Public","name":"CodeSage","owner":"amazon-science","isFork":false,"description":"CodeSage: Code Representation Learning At Scale (ICLR 2024)","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":0,"starsCount":74,"forksCount":3,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-28T11:58:45.654Z"}},{"type":"Public","name":"glass-text-spotting","owner":"amazon-science","isFork":false,"description":"Official implementation for \"GLASS: Global to Local Attention for Scene-Text Spotting\" (ECCV'22)","allTopics":["ocr","deep-learning","detection","attention","text-spotting"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":12,"starsCount":99,"forksCount":13,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-28T07:35:37.877Z"}},{"type":"Public","name":"chronos-forecasting","owner":"amazon-science","isFork":false,"description":"Chronos: Pretrained (Language) Models for Probabilistic Time Series Forecasting","allTopics":["machine-learning","timeseries","time-series","transformers","artificial-intelligence","forecasting","pretrained-models","time-series-forecasting","huggingface","huggingface-transformers","foundation-models","large-language-models","llm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":20,"starsCount":2080,"forksCount":245,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,1,3,4,2,2,0,0,6,5,1,5,4,2,1,5,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-27T21:11:05.412Z"}},{"type":"Public","name":"azcausal","owner":"amazon-science","isFork":false,"description":"Causal Inference in Python","allTopics":["panel","causal-inference","did","sdid"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":32,"forksCount":3,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-27T15:05:15.860Z"}},{"type":"Public","name":"MASSIVE-AMR","owner":"amazon-science","isFork":false,"description":"A dataset with more than 84,000 manually annotated AMR graphs for 1,685 information-seeking utterances mapped to 50+ typologically diverse languages.","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":0,"starsCount":3,"forksCount":1,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-25T19:31:46.199Z"}},{"type":"Public","name":"mezo_svrg","owner":"amazon-science","isFork":false,"description":"Code the ICML 2024 paper: \"Variance-reduced Zeroth-Order Methods for Fine-Tuning Language Models\"","allTopics":["machine-learning","deep-learning","optimization","machine-learning-algorithms","language-model","optimization-algorithms","variance-reduction","fine-tuning","svrg","zero-order-methods","large-language-models","llms","llm-training"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":4,"forksCount":0,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-25T15:50:07.075Z"}},{"type":"Public","name":"tabsyn","owner":"amazon-science","isFork":false,"description":"Official Implementations of \"Mixed-Type Tabular Data Synthesis with Score-based Diffusion in Latent Space\"\"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":9,"starsCount":59,"forksCount":25,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-21T00:35:33.985Z"}},{"type":"Public","name":"preference-driven-mt","owner":"amazon-science","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":0,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-20T21:41:26.236Z"}},{"type":"Public","name":"steerfair","owner":"amazon-science","isFork":false,"description":"Discovering Bias in Latent Space: An Unsupervised Debiasing Approach (ICML 2024)","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":0,"starsCount":3,"forksCount":1,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-20T19:52:38.265Z"}},{"type":"Public","name":"carbon-assessment-with-ml","owner":"amazon-science","isFork":false,"description":"CaML: Carbon Footprinting of Household Products with Zero-Shot Semantic Text Similarity","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":1,"starsCount":23,"forksCount":3,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-18T15:52:00.942Z"}}],"repositoryCount":295,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"amazon-science repositories"}