PlanGPT: Enhancing Urban Planning with Tailored Language Model and Efficient Retrieval (2402.19273v1)
Abstract: In the field of urban planning, general-purpose LLMs often struggle to meet the specific needs of planners. Tasks like generating urban planning texts, retrieving related information, and evaluating planning documents pose unique challenges. To enhance the efficiency of urban professionals and overcome these obstacles, we introduce PlanGPT, the first specialized LLM tailored for urban and spatial planning. Developed through collaborative efforts with institutions like the Chinese Academy of Urban Planning, PlanGPT leverages a customized local database retrieval framework, domain-specific fine-tuning of base models, and advanced tooling capabilities. Empirical tests demonstrate that PlanGPT has achieved advanced performance, delivering responses of superior quality precisely tailored to the intricacies of urban planning.
- Deepspeed-inference: enabling efficient inference of transformer models at unprecedented scale. In SC22: International Conference for High Performance Computing, Networking, Storage and Analysis, pages 1–15. IEEE.
- Anthropic. 2023. Model card and evaluations for claude models.
- Self-rag: Learning to retrieve, generate, and critique through self-reflection. arXiv preprint arXiv:2310.11511.
- Baichuan. 2023. Baichuan 2: Open large-scale language models. arXiv preprint arXiv:2309.10305.
- Improving language models by retrieving from trillions of tokens. In International conference on machine learning, pages 2206–2240. PMLR.
- A large annotated corpus for learning natural language inference. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing (EMNLP). Association for Computational Linguistics.
- SemEval-2017 task 1: Semantic textual similarity multilingual and crosslingual focused evaluation. In Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval-2017), pages 1–14, Vancouver, Canada. Association for Computational Linguistics.
- Bge m3-embedding: Multi-lingual, multi-functionality, multi-granularity text embeddings through self-knowledge distillation.
- Lift yourself up: Retrieval-augmented text generation with self-memory. Advances in Neural Information Processing Systems, 36.
- Vicuna: An open-source chatbot impressing gpt-4 with 90%* chatgpt quality.
- Deep reinforcement learning from human preferences.
- Chatlaw. https://github.com/PKU-YuanGroup/ChatLaw.
- Pre-training with whole word masking for chinese bert.
- Lert: A linguistically-motivated pre-trained language model.
- Efficient and effective text encoding for chinese llama and alpaca. arXiv preprint arXiv:2304.08177.
- Tri Dao. 2023. Flashattention-2: Faster attention with better parallelism and work partitioning. arXiv preprint arXiv:2307.08691.
- Google DeepMind. 2023a. Bard. https://bard.google.com.
- Google DeepMind. 2023b. Gemini. https://gemini.google.com.
- Mods: Model-oriented data selection for instruction tuning. arXiv preprint arXiv:2311.15653.
- Glm: General language model pretraining with autoregressive blank infilling. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 320–335.
- Jakubik et al. 2023a. Prithvi-100M.
- Rohan Anil et al. 2023b. Palm 2 technical report.
- A framework for few-shot language model evaluation.
- Simcse: Simple contrastive learning of sentence embeddings.
- HIT-SCIR. 2024. Chinese-mixtral-8x7b: An open-source mixture-of-experts llm. https://github.com/HIT-SCIR/Chinese-Mixtral-8x7B.
- Metagpt: Meta programming for multi-agent collaborative framework. arXiv preprint arXiv:2308.00352.
- Raven: In-context learning with retrieval augmented encoder-decoder language models. arXiv preprint arXiv:2308.07922.
- C-eval: A multi-level multi-discipline chinese evaluation suite for foundation models. In Advances in Neural Information Processing Systems.
- Active retrieval augmented generation. arXiv preprint arXiv:2305.06983.
- Tree of clarifications: Answering ambiguous questions with retrieval-augmented large language models. arXiv preprint arXiv:2310.14696.
- Efficient memory management for large language model serving with pagedattention.
- Cmmlu: Measuring massive multitask language understanding in chinese.
- From quantity to quality: Boosting llm performance with self-guided data selection for instruction tuning. ArXiv, abs/2308.12032.
- Self-alignment with instruction backtranslation. arXiv preprint arXiv:2308.06259.
- Ra-dit: Retrieval-augmented dual instruction tuning. arXiv preprint arXiv:2310.01352.
- C. Liu and W. Zhang. 2023. Social and spatial heterogeneities in covid-19 impacts on individual’s metro use: A big-data driven causality inference. Applied Geography, 155:102947.
- What makes good data for alignment? a comprehensive study of automatic data selection in instruction tuning. In The Twelfth International Conference on Learning Representations.
- Webglm: Towards an efficient web-enhanced question answering system with human preferences.
- LCQMC:a large-scale Chinese question matching corpus. In Proceedings of the 27th International Conference on Computational Linguistics, pages 1952–1962, Santa Fe, New Mexico, USA. Association for Computational Linguistics.
- Ilya Loshchilov and Frank Hutter. 2019. Decoupled weight decay regularization.
- Muffin: Curating multi-faceted instructions for improving instruction following. In The Twelfth International Conference on Learning Representations.
- Mobilityagent. https://github.com/XiaoLeGG/mobility-agent.
- Efficient estimation of word representations in vector space.
- Mistral-AI. 2023. mistral. https://mistral.ai/.
- Yohei Nakajima. Babyagi, 2023. URL https://github. com/yoheinakajima/babyagi. GitHub repository.
- Webgpt: Browser-assisted question-answering with human feedback. arXiv preprint arXiv:2112.09332.
- Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748.
- OpenAI. 2022. Chatgpt. https://chat.openai.com.
- OpenAI. 2023. Gpt-4 technical report.
- Wang Peng. 2023. Duomo/transgpt.
- Direct preference optimization: Your language model is secretly a reward model. Advances in Neural Information Processing Systems, 36.
- Nils Reimers and Iryna Gurevych. 2019. Sentence-bert: Sentence embeddings using siamese bert-networks.
- Ozan Sener and Silvio Savarese. 2017. Active learning for convolutional neural networks: A core-set approach. arXiv preprint arXiv:1708.00489.
- Threshold and moderating effects of land use on metro ridership in shenzhen: Implications for tod planning. Journal of Transport Geography, 89:102878.
- Built environment interventions for emission mitigation: A machine learning analysis of travel-related co2 in a developing city. Journal of Transport Geography, 110:103632.
- Significant Gravitas. AutoGPT.
- Charles Spearman. 1961. The proof and measurement of association between two things.
- Jianlin Su. 2022. Cosent.
- Recitation-augmented language models. arXiv preprint arXiv:2210.01296.
- Stanford alpaca: An instruction-following llama model. https://github.com/tatsu-lab/stanford_alpaca.
- Lagent Developer Team. 2023a. Lagent: InternLM a lightweight open-source framework that allows users to efficiently build large language model(llm)-based agents. https://github.com/InternLM/lagent.
- XAgent Team. 2023b. Xagent: An autonomous agent for complex task solving.
- Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971.
- Krish Mangroila Tycho Young, Andy Zhang. 2023. Mathgpt - an exploration into the field of mathematics with large language models.
- Laurens Van der Maaten and Geoffrey Hinton. 2008. Visualizing data using t-sne. Journal of machine learning research, 9(11).
- Huatuo: Tuning llama model with chinese medical knowledge. arXiv preprint arXiv:2304.06975.
- Self-instruct: Aligning language model with self generated instructions. arXiv preprint arXiv:2212.10560.
- Skywork: A more open bilingual foundation model.
- Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 38–45, Online. Association for Computational Linguistics.
- Autogen: Enabling next-gen llm applications via multi-agent conversation framework.
- Openagents: An open platform for language agents in the wild. arXiv preprint arXiv:2310.10634.
- Doctorglm: Fine-tuning your chinese doctor is not a herculean task. arXiv preprint arXiv:2304.01097.
- Wizardlm: Empowering large language models to follow complex instructions. arXiv preprint arXiv:2304.12244.
- PAWS-X: A cross-lingual adversarial dataset for paraphrase identification. CoRR, abs/1908.11828.
- Generate rather than retrieve: Large language models are strong context generators. arXiv preprint arXiv:2209.10063.
- Trafficgpt: Viewing, processing and interacting with traffic foundation models.
- Measuring megaregional structure in the pearl river delta by mobile phone signaling data: A complex network approach. Cities, 104:102809.
- Incorporating polycentric development and neighborhood life-circle planning for reducing driving in beijing: Nonlinear and threshold analysis. Cities, 121:103488.
- W. Zhang and K. Ning. 2023. Spatiotemporal heterogeneities in the causal effects of mobility intervention policies during the covid-19 outbreak: A spatially interrupted time-series (sits) analysis. Annals of the American Association of Geographers, 113(5):1112–1134.
- Earthgpt: A universal multi-modal large language model for multi-sensor image comprehension in remote sensing domain. arXiv preprint arXiv:2401.16822.
- Xuanyuan 2.0: A large chinese financial chat model with hundreds of billions parameters.
- Kun: Answer polishment for chinese self-alignment with instruction back-translation. arXiv preprint arXiv:2401.06477.
- Lima: Less is more for alignment. Advances in Neural Information Processing Systems, 36.
Collections
Sign up for free to add this paper to one or more collections.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.